[ 558.354899] env[62814]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62814) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 558.355334] env[62814]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62814) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 558.355334] env[62814]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62814) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 558.355663] env[62814]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 558.453787] env[62814]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62814) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:349}} [ 558.463501] env[62814]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62814) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:372}} [ 558.505740] env[62814]: INFO oslo_service.periodic_task [-] Skipping periodic task _heal_instance_info_cache because its interval is negative [ 559.067549] env[62814]: INFO nova.virt.driver [None req-50dc1253-da80-4157-9623-1e601aa5f6ac None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 559.136995] env[62814]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 559.137184] env[62814]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 559.137275] env[62814]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62814) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 562.027455] env[62814]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-d99aaaf4-b5ae-46a9-9197-b10a15c31d75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.044043] env[62814]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62814) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 562.044212] env[62814]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-8e1d0b66-6a2d-46b7-8c2d-a1312cb2d73e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.075749] env[62814]: INFO oslo_vmware.api [-] Successfully established new session; session ID is aa192. [ 562.075904] env[62814]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 2.939s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 562.076445] env[62814]: INFO nova.virt.vmwareapi.driver [None req-50dc1253-da80-4157-9623-1e601aa5f6ac None None] VMware vCenter version: 7.0.3 [ 562.079868] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657bf71b-d187-4ba3-946e-8ce5b17e8b9c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.100599] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad665fe-ae4f-4050-b0df-64350bfc1ff4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.106363] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee80d659-ffa6-444f-b686-5b6adef55205 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.112993] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd35f85-09e9-4dc5-b90c-8fd36dcc44d1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.125924] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b6b618-cb50-4644-98d2-8b056d33530a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.131791] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63524ea6-5be3-486f-b4e1-d9d03dcd003c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.162315] env[62814]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-3e1cdff6-b639-4385-982d-44ea4bfd82f7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.167387] env[62814]: DEBUG nova.virt.vmwareapi.driver [None req-50dc1253-da80-4157-9623-1e601aa5f6ac None None] Extension org.openstack.compute already exists. {{(pid=62814) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 562.169961] env[62814]: INFO nova.compute.provider_config [None req-50dc1253-da80-4157-9623-1e601aa5f6ac None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 562.673312] env[62814]: DEBUG nova.context [None req-50dc1253-da80-4157-9623-1e601aa5f6ac None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),5919660c-8efa-42fd-bda5-167c43697677(cell1) {{(pid=62814) load_cells /opt/stack/nova/nova/context.py:464}} [ 562.675407] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 562.675644] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 562.676358] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 562.676776] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Acquiring lock "5919660c-8efa-42fd-bda5-167c43697677" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 562.676997] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Lock "5919660c-8efa-42fd-bda5-167c43697677" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 562.678045] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Lock "5919660c-8efa-42fd-bda5-167c43697677" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 562.697774] env[62814]: INFO dbcounter [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Registered counter for database nova_cell0 [ 562.705963] env[62814]: INFO dbcounter [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Registered counter for database nova_cell1 [ 563.150411] env[62814]: DEBUG oslo_db.sqlalchemy.engines [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62814) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 563.152327] env[62814]: DEBUG oslo_db.sqlalchemy.engines [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62814) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 563.156058] env[62814]: ERROR nova.db.main.api [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 272, in main [ 563.156058] env[62814]: result = function(*args, **kwargs) [ 563.156058] env[62814]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.156058] env[62814]: return func(*args, **kwargs) [ 563.156058] env[62814]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 563.156058] env[62814]: result = fn(*args, **kwargs) [ 563.156058] env[62814]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 563.156058] env[62814]: return f(*args, **kwargs) [ 563.156058] env[62814]: File "/opt/stack/nova/nova/objects/service.py", line 557, in _db_service_get_minimum_version [ 563.156058] env[62814]: return db.service_get_minimum_version(context, binaries) [ 563.156058] env[62814]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 563.156058] env[62814]: _check_db_access() [ 563.156058] env[62814]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 563.156058] env[62814]: stacktrace = ''.join(traceback.format_stack()) [ 563.156058] env[62814]: [ 563.157247] env[62814]: ERROR nova.db.main.api [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 272, in main [ 563.157247] env[62814]: result = function(*args, **kwargs) [ 563.157247] env[62814]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 563.157247] env[62814]: return func(*args, **kwargs) [ 563.157247] env[62814]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 563.157247] env[62814]: result = fn(*args, **kwargs) [ 563.157247] env[62814]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 563.157247] env[62814]: return f(*args, **kwargs) [ 563.157247] env[62814]: File "/opt/stack/nova/nova/objects/service.py", line 557, in _db_service_get_minimum_version [ 563.157247] env[62814]: return db.service_get_minimum_version(context, binaries) [ 563.157247] env[62814]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 563.157247] env[62814]: _check_db_access() [ 563.157247] env[62814]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 563.157247] env[62814]: stacktrace = ''.join(traceback.format_stack()) [ 563.157247] env[62814]: [ 563.157667] env[62814]: WARNING nova.objects.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Failed to get minimum service version for cell 5919660c-8efa-42fd-bda5-167c43697677 [ 563.157870] env[62814]: WARNING nova.objects.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 563.158254] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Acquiring lock "singleton_lock" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.158412] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Acquired lock "singleton_lock" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 563.158652] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Releasing lock "singleton_lock" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 563.158980] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Full set of CONF: {{(pid=62814) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/service.py:357}} [ 563.159136] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ******************************************************************************** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2804}} [ 563.159284] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] Configuration options gathered from: {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2805}} [ 563.159431] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 563.159624] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 563.159750] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ================================================================================ {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 563.159956] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] allow_resize_to_same_host = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.160141] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] arq_binding_timeout = 300 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.160274] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] backdoor_port = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.160401] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] backdoor_socket = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.160568] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] block_device_allocate_retries = 60 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.160729] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] block_device_allocate_retries_interval = 3 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.161142] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cert = self.pem {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.161334] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.161509] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute_monitors = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.161679] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] config_dir = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.161983] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] config_drive_format = iso9660 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.161983] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.162160] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] config_source = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.162328] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] console_host = devstack {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.162491] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] control_exchange = nova {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.162649] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cpu_allocation_ratio = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.162806] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] daemon = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.162972] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] debug = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.163142] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] default_access_ip_network_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.163384] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] default_availability_zone = nova {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.163463] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] default_ephemeral_format = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.163607] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] default_green_pool_size = 1000 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.163841] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.164010] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] default_schedule_zone = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.164175] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] disk_allocation_ratio = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.164332] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] enable_new_services = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.164508] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] enabled_apis = ['osapi_compute'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.164668] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] enabled_ssl_apis = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.164826] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] flat_injected = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.164983] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] force_config_drive = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.165158] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] force_raw_images = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.165325] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] graceful_shutdown_timeout = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.165482] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] heal_instance_info_cache_interval = -1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.165703] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] host = cpu-1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.165878] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.166050] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.166215] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.166422] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.166583] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] instance_build_timeout = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.166738] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] instance_delete_interval = 300 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.166928] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] instance_format = [instance: %(uuid)s] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.167103] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] instance_name_template = instance-%08x {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.167314] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] instance_usage_audit = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.167493] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] instance_usage_audit_period = month {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.167658] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.167821] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.167987] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] internal_service_availability_zone = internal {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.168185] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] key = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.168349] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] live_migration_retry_count = 30 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.168515] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] log_color = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.168677] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] log_config_append = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.168839] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.168996] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] log_dir = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.169167] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] log_file = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.169318] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] log_options = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.169484] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] log_rotate_interval = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.169650] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] log_rotate_interval_type = days {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.169812] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] log_rotation_type = none {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.169938] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.170074] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.170241] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.170401] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.170524] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.170679] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] long_rpc_timeout = 1800 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.170832] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] max_concurrent_builds = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.170985] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] max_concurrent_live_migrations = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.171154] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] max_concurrent_snapshots = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.171310] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] max_local_block_devices = 3 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.171465] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] max_logfile_count = 30 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.171620] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] max_logfile_size_mb = 200 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.171773] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] maximum_instance_delete_attempts = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.171935] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] metadata_listen = 0.0.0.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.172111] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] metadata_listen_port = 8775 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.172279] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] metadata_workers = 2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.172436] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] migrate_max_retries = -1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.172597] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] mkisofs_cmd = genisoimage {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.172801] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.172930] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] my_ip = 10.180.1.21 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.173182] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.173361] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] network_allocate_retries = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.173546] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.173710] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.173872] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] osapi_compute_listen_port = 8774 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.174051] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] osapi_compute_unique_server_name_scope = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.174230] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] osapi_compute_workers = 2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.174391] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] password_length = 12 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.174548] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] periodic_enable = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.174705] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] periodic_fuzzy_delay = 60 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.174873] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] pointer_model = usbtablet {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.175043] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] preallocate_images = none {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.175208] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] publish_errors = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.175340] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] pybasedir = /opt/stack/nova {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.175531] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ram_allocation_ratio = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.175703] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] rate_limit_burst = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.175871] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] rate_limit_except_level = CRITICAL {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.176041] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] rate_limit_interval = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.176210] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] reboot_timeout = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.176365] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] reclaim_instance_interval = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.176516] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] record = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.176680] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] reimage_timeout_per_gb = 60 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.176843] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] report_interval = 120 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.177037] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] rescue_timeout = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.177211] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] reserved_host_cpus = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.177369] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] reserved_host_disk_mb = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.177526] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] reserved_host_memory_mb = 512 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.177683] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] reserved_huge_pages = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.177841] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] resize_confirm_window = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.178019] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] resize_fs_using_block_device = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.178202] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] resume_guests_state_on_host_boot = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.178374] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.178533] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] rpc_response_timeout = 60 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.178693] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] run_external_periodic_tasks = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.178858] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] running_deleted_instance_action = reap {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.179026] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.179191] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] running_deleted_instance_timeout = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.179381] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler_instance_sync_interval = 120 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.179548] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_down_time = 720 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.179712] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] servicegroup_driver = db {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.179866] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] shell_completion = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.180034] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] shelved_offload_time = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.180200] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] shelved_poll_interval = 3600 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.180366] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] shutdown_timeout = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.180525] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] source_is_ipv6 = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.180682] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ssl_only = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.180927] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.181110] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] sync_power_state_interval = 600 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.181271] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] sync_power_state_pool_size = 1000 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.181436] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] syslog_log_facility = LOG_USER {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.181591] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] tempdir = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.181748] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] timeout_nbd = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.181912] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] transport_url = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.182084] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] update_resources_interval = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.182247] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] use_cow_images = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.182403] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] use_journal = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.182556] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] use_json = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.182709] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] use_rootwrap_daemon = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.182863] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] use_stderr = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.183026] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] use_syslog = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.183185] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vcpu_pin_set = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.183348] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plugging_is_fatal = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.183518] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plugging_timeout = 300 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.183677] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] virt_mkfs = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.183832] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] volume_usage_poll_interval = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.183988] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] watch_log_file = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.184167] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] web = /usr/share/spice-html5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 563.184347] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.184512] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.184670] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.184836] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_concurrency.disable_process_locking = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.185153] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.185358] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.185529] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.185700] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.185867] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.186046] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.186243] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.auth_strategy = keystone {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.186409] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.compute_link_prefix = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.186579] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.186749] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.dhcp_domain = novalocal {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.186927] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.enable_instance_password = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.187118] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.glance_link_prefix = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.187286] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.187454] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.187615] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.instance_list_per_project_cells = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.187776] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.list_records_by_skipping_down_cells = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.187937] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.local_metadata_per_cell = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.188146] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.max_limit = 1000 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.188325] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.metadata_cache_expiration = 15 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.188502] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.neutron_default_tenant_id = default {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.188671] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.response_validation = warn {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.188841] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.use_neutron_default_nets = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.189016] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.189190] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.189402] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.189582] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.189755] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.vendordata_dynamic_targets = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.189920] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.vendordata_jsonfile_path = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.190115] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.190307] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.backend = dogpile.cache.memcached {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.190473] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.backend_argument = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.190633] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.backend_expiration_time = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.190802] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.config_prefix = cache.oslo {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.190968] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.dead_timeout = 60.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.191147] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.debug_cache_backend = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.191310] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.enable_retry_client = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.191469] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.enable_socket_keepalive = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.191635] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.enabled = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.191796] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.enforce_fips_mode = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.191959] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.expiration_time = 600 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.192133] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.hashclient_retry_attempts = 2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.192298] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.192458] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.memcache_dead_retry = 300 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.192615] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.memcache_password = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.192775] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.192935] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.193109] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.memcache_pool_maxsize = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.193273] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.193435] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.memcache_sasl_enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.193611] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.193777] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.193933] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.memcache_username = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.194111] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.proxies = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.194278] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.redis_db = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.194434] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.redis_password = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.194603] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.194778] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.194945] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.redis_server = localhost:6379 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.195125] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.redis_socket_timeout = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.195285] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.redis_username = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.195446] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.retry_attempts = 2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.195609] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.retry_delay = 0.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.195768] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.socket_keepalive_count = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.195929] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.socket_keepalive_idle = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.196100] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.socket_keepalive_interval = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.196260] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.tls_allowed_ciphers = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.196455] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.tls_cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.196628] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.tls_certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.196791] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.tls_enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.196979] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cache.tls_keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.197177] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.auth_section = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.197354] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.auth_type = password {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.197516] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.197690] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.197849] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.198049] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.198222] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.cross_az_attach = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.198386] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.debug = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.198545] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.endpoint_template = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.198706] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.http_retries = 3 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.198867] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.199036] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.199215] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.os_region_name = RegionOne {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.199379] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.199535] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cinder.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.199701] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.199858] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.cpu_dedicated_set = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.200023] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.cpu_shared_set = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.200191] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.image_type_exclude_list = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.200351] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.200509] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.200668] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.200826] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.200990] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.201190] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.resource_provider_association_refresh = 300 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.201350] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.201507] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.shutdown_retry_interval = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.201683] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.201859] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] conductor.workers = 2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.202048] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] console.allowed_origins = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.202213] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] console.ssl_ciphers = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.202383] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] console.ssl_minimum_version = default {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.202549] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] consoleauth.enforce_session_timeout = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.202713] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] consoleauth.token_ttl = 600 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.202881] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.203048] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.203217] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.203374] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.connect_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.203530] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.connect_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.203683] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.endpoint_override = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.203852] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.203995] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.204165] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.max_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.204319] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.min_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.204473] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.region_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.204625] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.retriable_status_codes = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.204781] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.service_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.204947] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.service_type = accelerator {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.205122] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.205281] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.status_code_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.205435] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.status_code_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.205587] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.205760] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.205919] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] cyborg.version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.206101] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.asyncio_connection = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.206265] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.asyncio_slave_connection = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.206434] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.backend = sqlalchemy {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.206603] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.connection = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.206765] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.connection_debug = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.206956] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.connection_parameters = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.207151] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.connection_recycle_time = 3600 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.207317] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.connection_trace = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.207481] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.db_inc_retry_interval = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.207647] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.db_max_retries = 20 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.207807] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.db_max_retry_interval = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.208013] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.db_retry_interval = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.208211] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.max_overflow = 50 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.208377] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.max_pool_size = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.208537] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.max_retries = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.208706] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.208862] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.mysql_wsrep_sync_wait = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.209028] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.pool_timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.209196] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.retry_interval = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.209355] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.slave_connection = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.209515] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.sqlite_synchronous = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.209678] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] database.use_db_reconnect = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.209843] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.asyncio_connection = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.210006] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.asyncio_slave_connection = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.210186] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.backend = sqlalchemy {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.210353] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.connection = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.210514] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.connection_debug = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.210687] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.connection_parameters = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.210850] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.connection_recycle_time = 3600 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.211017] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.connection_trace = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.211182] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.db_inc_retry_interval = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.211346] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.db_max_retries = 20 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.211507] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.db_max_retry_interval = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.211667] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.db_retry_interval = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.211826] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.max_overflow = 50 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.211986] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.max_pool_size = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.212160] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.max_retries = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.212329] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.212486] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.212643] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.pool_timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.212801] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.retry_interval = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.212956] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.slave_connection = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.213130] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] api_database.sqlite_synchronous = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.213305] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] devices.enabled_mdev_types = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.213479] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.213648] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.213809] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ephemeral_storage_encryption.enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.213972] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.214155] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.api_servers = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.214320] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.214480] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.214642] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.214800] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.connect_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.214959] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.connect_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.215136] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.debug = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.215303] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.default_trusted_certificate_ids = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.215465] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.enable_certificate_validation = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.215625] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.enable_rbd_download = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.215782] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.endpoint_override = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.215947] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.216122] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.216283] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.max_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.216437] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.min_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.216598] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.num_retries = 3 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.216763] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.rbd_ceph_conf = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.216961] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.rbd_connect_timeout = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.217138] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.rbd_pool = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.217308] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.rbd_user = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.217476] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.region_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.217634] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.retriable_status_codes = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.217791] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.service_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.217961] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.service_type = image {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.218162] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.218325] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.status_code_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.218482] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.status_code_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.218640] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.218817] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.218982] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.verify_glance_signatures = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.219155] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] glance.version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.219324] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] guestfs.debug = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.219488] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.auth_section = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.219647] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.auth_type = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.219803] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.219958] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.220135] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.220293] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.connect_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.220448] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.connect_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.220601] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.endpoint_override = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.220758] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.220912] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.221080] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.max_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.221236] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.min_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.221390] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.region_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.221543] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.retriable_status_codes = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.221698] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.service_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.221861] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.service_type = shared-file-system {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.222030] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.share_apply_policy_timeout = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.222197] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.222351] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.status_code_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.222505] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.status_code_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.222657] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.222833] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.222991] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] manila.version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.223174] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] mks.enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.223520] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.223708] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] image_cache.manager_interval = 2400 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.223875] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] image_cache.precache_concurrency = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.224053] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] image_cache.remove_unused_base_images = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.224224] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.224452] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.224653] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] image_cache.subdirectory_name = _base {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.224830] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.api_max_retries = 60 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.224995] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.api_retry_interval = 2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.225171] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.auth_section = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.225332] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.auth_type = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.225489] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.225645] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.225809] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.225971] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.conductor_group = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.226145] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.connect_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.226305] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.connect_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.226458] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.endpoint_override = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.226619] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.226774] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.226953] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.max_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.227142] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.min_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.227311] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.peer_list = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.227469] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.region_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.227624] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.retriable_status_codes = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.227785] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.serial_console_state_timeout = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.227964] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.service_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.228170] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.service_type = baremetal {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.228332] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.shard = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.228495] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.228695] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.status_code_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.228809] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.status_code_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.228966] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.229162] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.229324] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ironic.version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.229505] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.229679] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] key_manager.fixed_key = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.229857] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.230036] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.barbican_api_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.230212] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.barbican_endpoint = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.230384] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.barbican_endpoint_type = public {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.230573] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.barbican_region_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.230696] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.230852] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.231022] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.231189] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.231345] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.231506] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.number_of_retries = 60 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.231666] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.retry_delay = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.231826] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.send_service_user_token = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.231986] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.232160] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.232319] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.verify_ssl = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.232475] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican.verify_ssl_path = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.232636] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican_service_user.auth_section = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.232797] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican_service_user.auth_type = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.232951] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican_service_user.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.233120] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican_service_user.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.233282] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican_service_user.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.233443] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican_service_user.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.233598] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican_service_user.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.233758] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican_service_user.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.233912] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] barbican_service_user.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.234096] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.approle_role_id = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.234251] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.approle_secret_id = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.234418] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.kv_mountpoint = secret {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.234575] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.kv_path = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.234735] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.kv_version = 2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.234895] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.namespace = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.235065] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.root_token_id = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.235223] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.ssl_ca_crt_file = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.235385] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.timeout = 60.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.235546] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.use_ssl = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.235714] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.235880] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.236054] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.236221] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.236378] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.connect_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.236535] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.connect_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.236690] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.endpoint_override = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.236850] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.237039] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.237205] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.max_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.237362] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.min_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.237521] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.region_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.237680] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.retriable_status_codes = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.237836] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.service_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.238038] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.service_type = identity {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.238213] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.238372] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.status_code_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.238528] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.status_code_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.238684] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.238862] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.239030] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] keystone.version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.239226] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.ceph_mount_options = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.239617] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.239802] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.connection_uri = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.239968] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.cpu_mode = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.240153] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.240323] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.cpu_models = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.240492] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.cpu_power_governor_high = performance {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.240660] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.240821] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.cpu_power_management = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.240990] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.241174] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.device_detach_attempts = 8 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.241338] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.device_detach_timeout = 20 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.241503] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.disk_cachemodes = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.241662] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.disk_prefix = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.241822] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.enabled_perf_events = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.241983] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.file_backed_memory = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.242162] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.gid_maps = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.242318] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.hw_disk_discard = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.242472] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.hw_machine_type = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.242636] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.images_rbd_ceph_conf = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.242796] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.242954] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.243135] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.images_rbd_glance_store_name = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.243302] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.images_rbd_pool = rbd {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.243469] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.images_type = default {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.243625] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.images_volume_group = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.243783] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.inject_key = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.243941] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.inject_partition = -2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.244152] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.inject_password = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.244290] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.iscsi_iface = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.244433] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.iser_use_multipath = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.244592] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.244751] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.244913] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_downtime = 500 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.245084] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.245247] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.245403] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_inbound_addr = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.245563] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.245722] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.245877] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_scheme = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.246055] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_timeout_action = abort {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.246222] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_tunnelled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.246379] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_uri = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.246538] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.live_migration_with_native_tls = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.246693] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.max_queues = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.246853] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.247118] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.247290] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.nfs_mount_options = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.247578] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.247753] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.247947] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.248147] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.248314] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.248475] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.num_pcie_ports = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.248639] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.248801] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.pmem_namespaces = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.248960] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.quobyte_client_cfg = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.249256] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.249428] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.249591] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.249751] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.249907] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.rbd_secret_uuid = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.250073] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.rbd_user = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.250237] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.250405] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.250560] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.rescue_image_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.250715] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.rescue_kernel_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.250870] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.rescue_ramdisk_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.251047] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.251213] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.rx_queue_size = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.251378] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.smbfs_mount_options = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.251661] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.251833] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.snapshot_compression = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.251999] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.snapshot_image_format = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.252232] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.252397] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.sparse_logical_volumes = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.252558] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.swtpm_enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.252722] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.swtpm_group = tss {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.252885] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.swtpm_user = tss {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.253064] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.sysinfo_serial = unique {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.253222] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.tb_cache_size = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.253377] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.tx_queue_size = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.253535] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.uid_maps = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.253692] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.use_virtio_for_bridges = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.253856] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.virt_type = kvm {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.254031] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.volume_clear = zero {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.254204] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.volume_clear_size = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.254383] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.volume_enforce_multipath = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.254516] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.volume_use_multipath = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.254669] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.vzstorage_cache_path = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.254832] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.254996] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.255173] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.255339] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.255632] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.255811] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.vzstorage_mount_user = stack {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.255977] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.256168] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.auth_section = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.256342] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.auth_type = password {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.256503] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.256662] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.256822] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.257019] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.connect_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.257196] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.connect_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.257366] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.default_floating_pool = public {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.257525] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.endpoint_override = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.257686] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.extension_sync_interval = 600 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.257846] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.http_retries = 3 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.258057] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.258230] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.258392] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.max_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.258562] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.258722] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.min_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.258892] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.ovs_bridge = br-int {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.259076] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.physnets = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.259250] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.region_name = RegionOne {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.259410] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.retriable_status_codes = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.259579] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.service_metadata_proxy = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.259738] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.service_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.259904] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.service_type = network {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.260079] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.260243] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.status_code_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.260401] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.status_code_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.260559] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.260737] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.260900] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] neutron.version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.261087] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] notifications.bdms_in_notifications = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.261264] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] notifications.default_level = INFO {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.261428] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] notifications.include_share_mapping = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.261601] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] notifications.notification_format = unversioned {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.261762] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] notifications.notify_on_state_change = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.261935] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.262130] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] pci.alias = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.262336] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] pci.device_spec = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.262536] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] pci.report_in_placement = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.262733] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.auth_section = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.262914] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.auth_type = password {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.263101] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.263268] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.263453] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.263619] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.263776] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.connect_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.263935] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.connect_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.264109] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.default_domain_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.264266] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.default_domain_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.264467] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.domain_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.264576] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.domain_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.264731] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.endpoint_override = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.264891] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.265209] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.265268] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.max_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.265387] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.min_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.265553] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.password = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.265711] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.project_domain_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.265875] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.project_domain_name = Default {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.266078] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.project_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.266292] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.project_name = service {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.266468] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.region_name = RegionOne {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.266630] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.retriable_status_codes = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.266790] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.service_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.266981] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.service_type = placement {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.267171] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.267331] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.status_code_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.267486] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.status_code_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.267644] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.system_scope = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.267797] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.267972] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.trust_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.268156] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.user_domain_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.268324] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.user_domain_name = Default {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.268482] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.user_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.268649] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.username = nova {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.268828] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.268987] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] placement.version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.269185] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.cores = 20 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.269350] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.count_usage_from_placement = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.269518] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.269685] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.injected_file_content_bytes = 10240 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.269848] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.injected_file_path_length = 255 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.270021] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.injected_files = 5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.270194] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.instances = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.270359] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.key_pairs = 100 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.270521] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.metadata_items = 128 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.270682] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.ram = 51200 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.270843] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.recheck_quota = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.271025] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.server_group_members = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.271194] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.server_groups = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.271409] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.unified_limits_resource_list = ['servers'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.271583] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] quota.unified_limits_resource_strategy = require {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.271754] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.271917] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.272094] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler.image_metadata_prefilter = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.272258] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.272419] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler.max_attempts = 3 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.272578] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler.max_placement_results = 1000 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.272740] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.272900] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.273077] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.273249] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] scheduler.workers = 2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.273419] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.273586] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.273764] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.273932] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.274112] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.274365] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.274637] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.274944] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.275252] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.host_subset_size = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.275538] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.275817] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.276041] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.image_props_weight_multiplier = 0.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.276230] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.image_props_weight_setting = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.276415] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.276609] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.isolated_hosts = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.276799] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.isolated_images = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.277009] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.277189] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.277358] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.277522] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.pci_in_placement = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.277744] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.277985] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.278192] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.278362] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.278529] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.278692] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.278855] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.track_instance_changes = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.279047] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.279246] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] metrics.required = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.279389] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] metrics.weight_multiplier = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.279550] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.279715] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] metrics.weight_setting = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.280035] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.280216] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] serial_console.enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.280392] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] serial_console.port_range = 10000:20000 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.280561] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.280730] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.280897] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] serial_console.serialproxy_port = 6083 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.281077] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_user.auth_section = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.281255] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_user.auth_type = password {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.281414] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_user.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.281570] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_user.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.281729] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_user.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.281885] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_user.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.282051] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_user.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.282242] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_user.send_service_user_token = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.282417] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_user.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.282575] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] service_user.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.282743] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.agent_enabled = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.282905] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.283226] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.283430] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.283598] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.html5proxy_port = 6082 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.283759] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.image_compression = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.283916] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.jpeg_compression = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.284089] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.playback_compression = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.284255] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.require_secure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.284424] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.server_listen = 127.0.0.1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.284661] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.284976] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.spice_direct_proxy_base_url = http://127.0.0.1:13002/nova {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.285210] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.streaming_mode = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.285392] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] spice.zlib_compression = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.285561] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] upgrade_levels.baseapi = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.285774] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] upgrade_levels.compute = auto {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.285968] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] upgrade_levels.conductor = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.286169] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] upgrade_levels.scheduler = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.286341] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.286503] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.286659] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vendordata_dynamic_auth.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.286816] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vendordata_dynamic_auth.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.287011] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.287184] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vendordata_dynamic_auth.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.287342] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.287502] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.287658] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vendordata_dynamic_auth.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.287830] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.api_retry_count = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.287989] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.ca_file = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.288176] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.288343] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.cluster_name = testcl1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.288505] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.connection_pool_size = 10 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.288661] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.console_delay_seconds = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.288826] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.datastore_regex = ^datastore.* {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.289046] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.289241] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.host_password = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.289424] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.host_port = 443 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.289596] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.host_username = administrator@vsphere.local {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.289764] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.insecure = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.289926] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.integration_bridge = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.290101] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.maximum_objects = 100 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.290260] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.pbm_default_policy = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.290420] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.pbm_enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.290576] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.pbm_wsdl_location = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.290741] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.290896] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.serial_port_proxy_uri = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.291063] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.serial_port_service_uri = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.291229] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.task_poll_interval = 0.5 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.291398] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.use_linked_clone = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.291563] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.vnc_keymap = en-us {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.291725] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.vnc_port = 5900 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.291887] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vmware.vnc_port_total = 10000 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.292083] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vnc.auth_schemes = ['none'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.292260] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vnc.enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.292544] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.292726] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.292896] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vnc.novncproxy_port = 6080 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.293098] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vnc.server_listen = 127.0.0.1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.293282] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.293446] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vnc.vencrypt_ca_certs = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.293605] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vnc.vencrypt_client_cert = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.293763] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vnc.vencrypt_client_key = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.293939] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.294116] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.disable_deep_image_inspection = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.294280] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.294441] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.294598] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.294758] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.disable_rootwrap = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.294908] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.enable_numa_live_migration = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.295081] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.295244] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.295404] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.295698] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.libvirt_disable_apic = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.295873] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.296053] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.296224] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.296390] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.296553] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.296741] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.296932] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.297115] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.297284] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.297479] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.297675] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.297847] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.client_socket_timeout = 900 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.298026] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.default_pool_size = 1000 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.298201] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.keep_alive = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.298367] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.max_header_line = 16384 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.298528] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.298687] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.ssl_ca_file = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.298846] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.ssl_cert_file = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.299021] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.ssl_key_file = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.299191] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.tcp_keepidle = 600 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.299375] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.299544] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] zvm.ca_file = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.299704] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] zvm.cloud_connector_url = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.299982] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.300173] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] zvm.reachable_timeout = 300 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.300350] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.300521] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.300694] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.connection_string = messaging:// {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.300859] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.enabled = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.301034] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.es_doc_type = notification {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.301203] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.es_scroll_size = 10000 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.301366] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.es_scroll_time = 2m {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.301526] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.filter_error_trace = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.301688] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.hmac_keys = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.301850] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.sentinel_service_name = mymaster {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.302014] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.socket_timeout = 0.1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.302179] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.trace_requests = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.302338] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler.trace_sqlalchemy = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.302512] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler_jaeger.process_tags = {} {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.302669] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler_jaeger.service_name_prefix = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.302827] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] profiler_otlp.service_name_prefix = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.302989] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] remote_debug.host = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.303166] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] remote_debug.port = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.303340] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.303502] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.303663] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.303822] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.303981] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.304160] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.304322] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.304482] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.304641] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.304837] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.304960] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.305144] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.305311] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.305473] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.305633] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.kombu_reconnect_splay = 0.0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.305798] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.305958] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.306133] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.306303] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.306460] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.306619] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.306781] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.306970] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.307153] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.307318] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.307477] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.307636] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.307795] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.307962] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.308162] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.308332] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.ssl = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.308503] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.308669] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.308832] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.309006] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.309187] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.309392] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.309585] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.309755] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_notifications.retry = -1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.309933] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.310119] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.310294] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.auth_section = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.310457] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.auth_type = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.310615] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.cafile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.310769] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.certfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.310929] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.collect_timing = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.311099] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.connect_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.311259] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.connect_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.311414] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.endpoint_id = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.311584] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.endpoint_interface = publicURL {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.311741] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.endpoint_override = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.311898] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.endpoint_region_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.312068] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.endpoint_service_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.312353] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.endpoint_service_type = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.312415] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.insecure = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.312538] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.keyfile = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.312691] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.max_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.312842] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.min_version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.312992] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.region_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.313162] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.retriable_status_codes = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.313317] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.service_name = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.313472] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.service_type = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.313629] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.split_loggers = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.313782] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.status_code_retries = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.313938] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.status_code_retry_delay = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.314130] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.timeout = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.314310] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.valid_interfaces = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.314468] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_limit.version = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.314634] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_reports.file_event_handler = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.314802] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.314958] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] oslo_reports.log_dir = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.315141] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.315306] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.315518] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.315696] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.315862] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.316033] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.316214] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.316374] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_ovs_privileged.group = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.316530] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.316696] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.316858] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.317057] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] vif_plug_ovs_privileged.user = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.317240] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.317425] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.317598] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.317780] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.317953] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.318186] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.318366] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.318530] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.318709] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.318878] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_ovs.isolate_vif = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.319058] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.319231] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.319397] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.319564] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.319723] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] os_vif_ovs.per_port_bridge = False {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.319892] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] privsep_osbrick.capabilities = [21] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.320065] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] privsep_osbrick.group = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.320226] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] privsep_osbrick.helper_command = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.320389] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.320550] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.320705] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] privsep_osbrick.user = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.320876] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.321047] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] nova_sys_admin.group = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.321295] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] nova_sys_admin.helper_command = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.321499] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.321666] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.321826] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] nova_sys_admin.user = None {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 563.321956] env[62814]: DEBUG oslo_service.backend.eventlet.service [None req-e576efa2-419d-4372-bb96-dc20e5548321 None None] ******************************************************************************** {{(pid=62814) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2828}} [ 563.322366] env[62814]: INFO nova.service [-] Starting compute node (version 31.0.1) [ 563.825937] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Getting list of instances from cluster (obj){ [ 563.825937] env[62814]: value = "domain-c8" [ 563.825937] env[62814]: _type = "ClusterComputeResource" [ 563.825937] env[62814]: } {{(pid=62814) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 563.827367] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ed123e-6859-4340-a95c-a7f62a0453c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.836308] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Got total of 0 instances {{(pid=62814) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 563.836873] env[62814]: WARNING nova.virt.vmwareapi.driver [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 563.837374] env[62814]: INFO nova.virt.node [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Generated node identity 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 [ 563.837650] env[62814]: INFO nova.virt.node [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Wrote node identity 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 to /opt/stack/data/n-cpu-1/compute_id [ 564.340768] env[62814]: WARNING nova.compute.manager [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Compute nodes ['7136a6f7-3927-4aa1-a4e9-7fcbd9976745'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 565.349494] env[62814]: INFO nova.compute.manager [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 566.358454] env[62814]: WARNING nova.compute.manager [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 566.358809] env[62814]: DEBUG oslo_concurrency.lockutils [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 566.358982] env[62814]: DEBUG oslo_concurrency.lockutils [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 566.359177] env[62814]: DEBUG oslo_concurrency.lockutils [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 566.359374] env[62814]: DEBUG nova.compute.resource_tracker [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 566.360312] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe78991-f2b6-48e9-a1f0-c30042f315f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.368866] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88431b1f-430d-4294-96cd-6a81d1f393c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.383639] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-119b24ab-a50f-4bcd-99ff-ed046399f6d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.389855] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a896da2d-93b3-4b38-b260-a79fbd0f9fc3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.418326] env[62814]: DEBUG nova.compute.resource_tracker [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180671MB free_disk=146GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 566.418470] env[62814]: DEBUG oslo_concurrency.lockutils [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 566.418716] env[62814]: DEBUG oslo_concurrency.lockutils [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 566.921063] env[62814]: WARNING nova.compute.resource_tracker [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] No compute node record for cpu-1:7136a6f7-3927-4aa1-a4e9-7fcbd9976745: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 could not be found. [ 567.424246] env[62814]: INFO nova.compute.resource_tracker [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 [ 568.932689] env[62814]: DEBUG nova.compute.resource_tracker [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 568.933144] env[62814]: DEBUG nova.compute.resource_tracker [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=149GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 569.090446] env[62814]: INFO nova.scheduler.client.report [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] [req-567d8211-8468-47bd-92c7-dfe228fad072] Created resource provider record via placement API for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 569.107680] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92098bca-9125-4641-8b63-dbd103a04780 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.115141] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-933f1f13-b981-400a-9f67-1d9d7162b846 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.144956] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66123c24-1c38-45eb-8dd2-bc3e7565df32 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.152423] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1e76b7-0631-4e24-80cf-d3dc07b8283a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.165627] env[62814]: DEBUG nova.compute.provider_tree [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 569.703882] env[62814]: DEBUG nova.scheduler.client.report [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 569.704147] env[62814]: DEBUG nova.compute.provider_tree [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 0 to 1 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 569.704285] env[62814]: DEBUG nova.compute.provider_tree [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 569.754303] env[62814]: DEBUG nova.compute.provider_tree [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 1 to 2 during operation: update_traits {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 570.259394] env[62814]: DEBUG nova.compute.resource_tracker [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 570.259755] env[62814]: DEBUG oslo_concurrency.lockutils [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.841s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 570.259755] env[62814]: DEBUG nova.service [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Creating RPC server for service compute {{(pid=62814) start /opt/stack/nova/nova/service.py:186}} [ 570.273469] env[62814]: DEBUG nova.service [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] Join ServiceGroup membership for this service compute {{(pid=62814) start /opt/stack/nova/nova/service.py:203}} [ 570.273652] env[62814]: DEBUG nova.servicegroup.drivers.db [None req-105cf5a6-ca47-442a-95fb-e9f417ab1567 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62814) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 574.277054] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_power_states {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 574.780053] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Getting list of instances from cluster (obj){ [ 574.780053] env[62814]: value = "domain-c8" [ 574.780053] env[62814]: _type = "ClusterComputeResource" [ 574.780053] env[62814]: } {{(pid=62814) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 574.781309] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b146d81-eb64-48c3-848d-67f81a0aa40d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.789504] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Got total of 0 instances {{(pid=62814) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 574.789732] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 574.790025] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Getting list of instances from cluster (obj){ [ 574.790025] env[62814]: value = "domain-c8" [ 574.790025] env[62814]: _type = "ClusterComputeResource" [ 574.790025] env[62814]: } {{(pid=62814) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 574.790844] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0f43aa-2071-4dc9-9dc4-fa612f51131c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.798820] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Got total of 0 instances {{(pid=62814) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 609.354449] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquiring lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.354449] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 609.857926] env[62814]: DEBUG nova.compute.manager [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 610.086613] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquiring lock "37649169-66e6-4ef1-a4cd-c65b9831004a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 610.086613] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lock "37649169-66e6-4ef1-a4cd-c65b9831004a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 610.421995] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 610.422310] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 610.426699] env[62814]: INFO nova.compute.claims [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 610.589740] env[62814]: DEBUG nova.compute.manager [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 611.126856] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 611.154603] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquiring lock "591e4a92-7c03-4d70-885f-49a66032c97e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 611.154823] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lock "591e4a92-7c03-4d70-885f-49a66032c97e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 611.543174] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfb5ab2-1974-4d7b-bcee-45a7241731c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.552728] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f395288-b6fb-47db-afc9-6f137e0e1e8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.589662] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515ac25c-ca53-4603-a695-7d7b55dbf93c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.598982] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afa7930-fd57-4a3a-a80f-fc02ac404c8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.616458] env[62814]: DEBUG nova.compute.provider_tree [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.657276] env[62814]: DEBUG nova.compute.manager [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 611.787885] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquiring lock "66db614c-d101-409d-a5c3-9de38e5a6eff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 611.788297] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lock "66db614c-d101-409d-a5c3-9de38e5a6eff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 612.119618] env[62814]: DEBUG nova.scheduler.client.report [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 612.212249] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 612.233806] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "be02bc93-1278-4b3d-afa3-270c84585d1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 612.234201] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "be02bc93-1278-4b3d-afa3-270c84585d1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 612.294793] env[62814]: DEBUG nova.compute.manager [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 612.626121] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 612.626611] env[62814]: DEBUG nova.compute.manager [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 612.630579] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.505s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 612.632265] env[62814]: INFO nova.compute.claims [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.738658] env[62814]: DEBUG nova.compute.manager [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 612.835818] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 612.908966] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquiring lock "f5504590-823d-4c37-8226-26684f7957a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 612.909195] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lock "f5504590-823d-4c37-8226-26684f7957a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.119590] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.119590] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.138463] env[62814]: DEBUG nova.compute.utils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 613.141045] env[62814]: DEBUG nova.compute.manager [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 613.141337] env[62814]: DEBUG nova.network.neutron [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 613.270352] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.412232] env[62814]: DEBUG nova.compute.manager [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 613.627699] env[62814]: DEBUG nova.compute.manager [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 613.647913] env[62814]: DEBUG nova.compute.manager [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 613.713528] env[62814]: DEBUG nova.policy [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e221719837034bbf972c36aa33292b83', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4ab33b31c8846b39eb29f2e905d059c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 613.847998] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208d4b21-6e32-4e36-bd79-b3cc8b902791 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.856210] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10ef186f-7530-417a-a20a-20de49961427 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.890326] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ed8be6-4d2e-4d09-8cd2-f663e77d0c83 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.895442] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "6a592192-1b41-4be2-84a6-c3b76a4e5643" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 613.895703] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 613.903558] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c98f2a-97c3-4f11-bfa4-c1fd277032f7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.922122] env[62814]: DEBUG nova.compute.provider_tree [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.942714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 614.173300] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 614.400032] env[62814]: DEBUG nova.compute.manager [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 614.425082] env[62814]: DEBUG nova.scheduler.client.report [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 614.672978] env[62814]: DEBUG nova.compute.manager [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 614.716883] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 614.716883] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 614.716883] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 614.717307] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 614.717307] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 614.717307] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 614.717307] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 614.717307] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 614.718263] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 614.718263] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 614.718263] env[62814]: DEBUG nova.virt.hardware [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 614.719320] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2569aac4-2487-4774-968f-a293b20debf1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.730709] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f229291-e121-42bf-b55f-630b113078a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.750768] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f839bdff-9f5a-49cb-801a-6a4043597fc5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.935715] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 614.943938] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.313s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 614.944365] env[62814]: DEBUG nova.compute.manager [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 614.947809] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.737s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 614.950246] env[62814]: INFO nova.compute.claims [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.073010] env[62814]: DEBUG nova.network.neutron [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Successfully created port: a896bdd8-35a5-4767-a971-88307c117164 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 615.464582] env[62814]: DEBUG nova.compute.utils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 615.466582] env[62814]: DEBUG nova.compute.manager [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 615.468812] env[62814]: DEBUG nova.network.neutron [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 615.775228] env[62814]: DEBUG nova.policy [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ede8459566cf4704a7aa558b0a9f7939', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6a5b7b2f93434cbe9df49ccc5374928c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 615.969325] env[62814]: DEBUG nova.compute.manager [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 616.126497] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-379a378d-94be-41f1-a0e2-3b7dc5def645 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.138472] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e1e285-a4f1-447b-ab16-f7599d444121 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.174398] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42655b2-ab5e-4dcf-b0cf-f4fdb1777223 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.186440] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf96318-cffa-4c55-a108-96f97aaa3414 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.208947] env[62814]: DEBUG nova.compute.provider_tree [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.718581] env[62814]: DEBUG nova.scheduler.client.report [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 616.981124] env[62814]: DEBUG nova.compute.manager [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 617.028185] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 617.028185] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.028185] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 617.028391] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.028391] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 617.028391] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 617.028391] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 617.028391] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 617.028535] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 617.028535] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 617.028972] env[62814]: DEBUG nova.virt.hardware [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 617.030215] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8667631-d205-4d9d-b131-f6a12e8da497 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.044032] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33807a37-a4ac-4544-a0c0-af9552cde597 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.176133] env[62814]: DEBUG nova.network.neutron [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Successfully created port: 5dc7f23e-f7a5-49d3-8451-88a7f80d7c90 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 617.228165] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.277s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.228165] env[62814]: DEBUG nova.compute.manager [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 617.229787] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.395s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.231370] env[62814]: INFO nova.compute.claims [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 617.733024] env[62814]: DEBUG nova.compute.utils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 617.733024] env[62814]: DEBUG nova.compute.manager [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Not allocating networking since 'none' was specified. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 618.235022] env[62814]: DEBUG nova.compute.manager [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 618.410093] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a365218-e30f-41b8-bf17-9d68aa2e9a4f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.418699] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87886e9d-279a-49de-bf5f-53947cf07b64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.469391] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d88bfb7-374c-4d16-9db1-1613d7616a46 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.477262] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64087f74-bfb6-48b3-a4ee-13c5832e6e4f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.493917] env[62814]: DEBUG nova.compute.provider_tree [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.515119] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.515676] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.515892] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.516086] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.516289] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.516484] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.516670] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.516839] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 618.516985] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 618.944555] env[62814]: DEBUG nova.network.neutron [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Successfully updated port: a896bdd8-35a5-4767-a971-88307c117164 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 619.001024] env[62814]: DEBUG nova.scheduler.client.report [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 619.020822] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.248195] env[62814]: DEBUG nova.compute.manager [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 619.272280] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 619.272551] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.272994] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 619.272994] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.272994] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 619.273310] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 619.273567] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 619.273764] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 619.273971] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 619.274190] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 619.274442] env[62814]: DEBUG nova.virt.hardware [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 619.275383] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecef179b-aa9c-4752-b13d-cfbdaf2c2e9c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.287451] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e135f5c-79c7-41c9-b8e7-a1b71cb0af4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.302318] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 619.312951] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 619.312951] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be730fae-1ee3-469d-a9e0-63ad7cd4866b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.323738] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Created folder: OpenStack in parent group-v4. [ 619.324060] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Creating folder: Project (5bd7e0f4c8954754a199086397be05f9). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 619.324153] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0aab0b79-8b0b-488f-b8dc-1983baab92e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.333407] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Created folder: Project (5bd7e0f4c8954754a199086397be05f9) in parent group-v845547. [ 619.333407] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Creating folder: Instances. Parent ref: group-v845548. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 619.333563] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c2b18ee6-1d8c-40e4-8eee-c95362a02db2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.343847] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Created folder: Instances in parent group-v845548. [ 619.344109] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 619.344300] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 619.344498] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1701829-9b0f-43e6-b644-9bfe8bf50913 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.367021] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 619.367021] env[62814]: value = "task-4293173" [ 619.367021] env[62814]: _type = "Task" [ 619.367021] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.378152] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293173, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.449239] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquiring lock "refresh_cache-32f1bd48-3a2e-42fe-8033-85d3ad55affe" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.449393] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquired lock "refresh_cache-32f1bd48-3a2e-42fe-8033-85d3ad55affe" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 619.449544] env[62814]: DEBUG nova.network.neutron [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.506390] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 619.506637] env[62814]: DEBUG nova.compute.manager [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 619.510157] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.240s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.511685] env[62814]: INFO nova.compute.claims [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 619.881613] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293173, 'name': CreateVM_Task, 'duration_secs': 0.362049} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.881613] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 619.885926] env[62814]: DEBUG oslo_vmware.service [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ed05d89-d54d-49b4-9ddc-da39e53776ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.893717] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.893717] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 619.894425] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 619.894606] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14ad5b64-b3f8-4305-9267-33089cf7b2f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.899203] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 619.899203] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525e7e21-b1a0-d637-f1dc-c1dee6d6fce2" [ 619.899203] env[62814]: _type = "Task" [ 619.899203] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.907925] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525e7e21-b1a0-d637-f1dc-c1dee6d6fce2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.971180] env[62814]: DEBUG nova.compute.manager [req-4eac5de1-8fad-42c5-8b21-f7ebf9e3da08 req-6d2fd55d-7ee5-4c65-a21b-96a80f0e4cc2 service nova] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Received event network-vif-plugged-a896bdd8-35a5-4767-a971-88307c117164 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 619.971180] env[62814]: DEBUG oslo_concurrency.lockutils [req-4eac5de1-8fad-42c5-8b21-f7ebf9e3da08 req-6d2fd55d-7ee5-4c65-a21b-96a80f0e4cc2 service nova] Acquiring lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 619.971180] env[62814]: DEBUG oslo_concurrency.lockutils [req-4eac5de1-8fad-42c5-8b21-f7ebf9e3da08 req-6d2fd55d-7ee5-4c65-a21b-96a80f0e4cc2 service nova] Lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 619.971180] env[62814]: DEBUG oslo_concurrency.lockutils [req-4eac5de1-8fad-42c5-8b21-f7ebf9e3da08 req-6d2fd55d-7ee5-4c65-a21b-96a80f0e4cc2 service nova] Lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 619.972148] env[62814]: DEBUG nova.compute.manager [req-4eac5de1-8fad-42c5-8b21-f7ebf9e3da08 req-6d2fd55d-7ee5-4c65-a21b-96a80f0e4cc2 service nova] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] No waiting events found dispatching network-vif-plugged-a896bdd8-35a5-4767-a971-88307c117164 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 619.972748] env[62814]: WARNING nova.compute.manager [req-4eac5de1-8fad-42c5-8b21-f7ebf9e3da08 req-6d2fd55d-7ee5-4c65-a21b-96a80f0e4cc2 service nova] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Received unexpected event network-vif-plugged-a896bdd8-35a5-4767-a971-88307c117164 for instance with vm_state building and task_state spawning. [ 620.021663] env[62814]: DEBUG nova.compute.utils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 620.024328] env[62814]: DEBUG nova.compute.manager [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 620.024328] env[62814]: DEBUG nova.network.neutron [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 620.032067] env[62814]: DEBUG nova.network.neutron [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.244456] env[62814]: DEBUG nova.policy [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0118373852b6451e9a86e5185195cddf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5f8144abf6b549cb8246aa63420e625a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 620.272508] env[62814]: DEBUG nova.network.neutron [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Successfully updated port: 5dc7f23e-f7a5-49d3-8451-88a7f80d7c90 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 620.413190] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.413468] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 620.413722] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.414461] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 620.414796] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 620.415377] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b04c7d0-fbb6-4856-8f16-6fde514d2403 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.433743] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 620.435092] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 620.436038] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67681bb-9117-4d7d-a6ad-ad9b9d5c7974 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.442829] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35e35c11-8e72-4d35-8ce7-23be4aa15078 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.447962] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 620.447962] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52059178-2865-61e2-e2c8-1050fe8e2795" [ 620.447962] env[62814]: _type = "Task" [ 620.447962] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.456701] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52059178-2865-61e2-e2c8-1050fe8e2795, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.476533] env[62814]: DEBUG nova.network.neutron [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Updating instance_info_cache with network_info: [{"id": "a896bdd8-35a5-4767-a971-88307c117164", "address": "fa:16:3e:75:f5:8f", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.161", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa896bdd8-35", "ovs_interfaceid": "a896bdd8-35a5-4767-a971-88307c117164", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.536147] env[62814]: DEBUG nova.compute.manager [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 620.751912] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6127f578-84b9-4d3c-b91d-160138041e4f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.758122] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a209b7b-fb1a-451c-9390-7619d77d7e2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.791094] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquiring lock "refresh_cache-37649169-66e6-4ef1-a4cd-c65b9831004a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.791094] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquired lock "refresh_cache-37649169-66e6-4ef1-a4cd-c65b9831004a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 620.791094] env[62814]: DEBUG nova.network.neutron [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 620.794574] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fccf042-7910-4c5b-bb36-4e32a21c589a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.801910] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a393cc7-e1d8-464a-8012-59d2478f689b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.817822] env[62814]: DEBUG nova.compute.provider_tree [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.959905] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Preparing fetch location {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 620.960232] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Creating directory with path [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 620.961547] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-83bb8034-dffa-44d7-b1fb-5767ef0fc126 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.976131] env[62814]: DEBUG nova.network.neutron [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Successfully created port: 320f12a1-7803-4527-847e-8660c500ed93 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 620.981649] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Releasing lock "refresh_cache-32f1bd48-3a2e-42fe-8033-85d3ad55affe" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.982049] env[62814]: DEBUG nova.compute.manager [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Instance network_info: |[{"id": "a896bdd8-35a5-4767-a971-88307c117164", "address": "fa:16:3e:75:f5:8f", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.161", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa896bdd8-35", "ovs_interfaceid": "a896bdd8-35a5-4767-a971-88307c117164", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 620.982515] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:75:f5:8f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a896bdd8-35a5-4767-a971-88307c117164', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 620.995597] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Creating folder: Project (e4ab33b31c8846b39eb29f2e905d059c). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 620.997956] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-391ff5e5-8c9b-4dbc-bd3e-d8fe7f9555c7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.998713] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Created directory with path [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 620.998923] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Fetch image to [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22/tmp-sparse.vmdk {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 620.999115] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Downloading image file data e7d68632-de75-4206-8f75-4abd879e1d22 to [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22/tmp-sparse.vmdk on the data store datastore2 {{(pid=62814) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 620.999848] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68631d10-c175-443f-aa3e-8b952907d77e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.014257] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Created folder: Project (e4ab33b31c8846b39eb29f2e905d059c) in parent group-v845547. [ 621.014257] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Creating folder: Instances. Parent ref: group-v845551. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 621.014257] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22efa792-394d-420f-84cd-0717f16e02c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.019645] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e803aa-5d33-49be-babe-f10351d100c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.025248] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Created folder: Instances in parent group-v845551. [ 621.025375] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 621.026709] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 621.026946] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-76f9e97e-63d0-4b3b-92a4-66ce39a8d350 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.054465] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23675757-1539-4534-a6f0-fbe257aecb70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.059825] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 621.059825] env[62814]: value = "task-4293176" [ 621.059825] env[62814]: _type = "Task" [ 621.059825] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.092373] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9093b009-9c44-41fa-9be2-45631034d6b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.099556] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293176, 'name': CreateVM_Task} progress is 15%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.104674] env[62814]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-87de9d3c-a6b2-4842-9783-e693e8bda0c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.194373] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Downloading image file data e7d68632-de75-4206-8f75-4abd879e1d22 to the data store datastore2 {{(pid=62814) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 621.270289] env[62814]: DEBUG oslo_vmware.rw_handles [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62814) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 621.338946] env[62814]: DEBUG nova.scheduler.client.report [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 621.433783] env[62814]: DEBUG nova.network.neutron [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.561028] env[62814]: DEBUG nova.compute.manager [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 621.583385] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293176, 'name': CreateVM_Task, 'duration_secs': 0.317403} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.588555] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 621.600166] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 621.601477] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.601477] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 621.601477] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.601477] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 621.601477] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 621.601681] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 621.601681] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 621.601795] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 621.601976] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 621.602161] env[62814]: DEBUG nova.virt.hardware [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 621.603075] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261cd373-838e-45e5-a1fc-e4140e16b430 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.610065] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.610557] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.611275] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 621.614974] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f10f890-5870-4d64-b77f-9a01df61518a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.625165] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e99332f-afd0-4fe6-8772-aa8aad13d7c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.631025] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for the task: (returnval){ [ 621.631025] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5286980a-81cb-ba77-d7b6-8418498065f2" [ 621.631025] env[62814]: _type = "Task" [ 621.631025] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.666143] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 621.666410] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 621.666626] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.753368] env[62814]: DEBUG nova.network.neutron [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Updating instance_info_cache with network_info: [{"id": "5dc7f23e-f7a5-49d3-8451-88a7f80d7c90", "address": "fa:16:3e:cc:4b:b6", "network": {"id": "fe8eed21-168a-4072-b84a-df0ba6420f17", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1129990206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5b7b2f93434cbe9df49ccc5374928c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5dc7f23e-f7", "ovs_interfaceid": "5dc7f23e-f7a5-49d3-8451-88a7f80d7c90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.853067] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 621.853916] env[62814]: DEBUG nova.compute.manager [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 621.857212] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.915s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 621.859033] env[62814]: INFO nova.compute.claims [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.944635] env[62814]: DEBUG oslo_vmware.rw_handles [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Completed reading data from the image iterator. {{(pid=62814) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 621.944913] env[62814]: DEBUG oslo_vmware.rw_handles [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 622.030977] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "17bfe703-ff96-4cb7-8535-eed57e10d673" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 622.031223] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "17bfe703-ff96-4cb7-8535-eed57e10d673" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 622.089245] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Downloaded image file data e7d68632-de75-4206-8f75-4abd879e1d22 to vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22/tmp-sparse.vmdk on the data store datastore2 {{(pid=62814) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 622.091124] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Caching image {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 622.091225] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Copying Virtual Disk [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22/tmp-sparse.vmdk to [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 622.092050] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b26d2da6-495d-47ba-83dd-6f6fdda4185b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.100249] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 622.100249] env[62814]: value = "task-4293177" [ 622.100249] env[62814]: _type = "Task" [ 622.100249] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.120027] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293177, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.264116] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Releasing lock "refresh_cache-37649169-66e6-4ef1-a4cd-c65b9831004a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 622.264116] env[62814]: DEBUG nova.compute.manager [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Instance network_info: |[{"id": "5dc7f23e-f7a5-49d3-8451-88a7f80d7c90", "address": "fa:16:3e:cc:4b:b6", "network": {"id": "fe8eed21-168a-4072-b84a-df0ba6420f17", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1129990206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5b7b2f93434cbe9df49ccc5374928c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5dc7f23e-f7", "ovs_interfaceid": "5dc7f23e-f7a5-49d3-8451-88a7f80d7c90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 622.264522] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:4b:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '22b8c642-38ad-4c11-9051-145ab3bc54f2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5dc7f23e-f7a5-49d3-8451-88a7f80d7c90', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 622.274543] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Creating folder: Project (6a5b7b2f93434cbe9df49ccc5374928c). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 622.274707] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e569a4c9-2e18-4a22-8375-e849f99046d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.290134] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Created folder: Project (6a5b7b2f93434cbe9df49ccc5374928c) in parent group-v845547. [ 622.290134] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Creating folder: Instances. Parent ref: group-v845554. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 622.290331] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7e9322f-e534-4d52-bfb5-53a75b4963a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.304262] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Created folder: Instances in parent group-v845554. [ 622.304498] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 622.304687] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 622.304881] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0aa30573-6b52-4b9d-b132-6deb761da56e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.332628] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 622.332628] env[62814]: value = "task-4293180" [ 622.332628] env[62814]: _type = "Task" [ 622.332628] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.343408] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293180, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.359065] env[62814]: DEBUG nova.compute.utils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 622.361251] env[62814]: DEBUG nova.compute.manager [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 622.361251] env[62814]: DEBUG nova.network.neutron [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 622.468717] env[62814]: DEBUG nova.policy [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3cbeb7b7e7fc46ffac95e891184f56b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97d11f49b6ad4c82956efa254d8995ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 622.539205] env[62814]: DEBUG nova.compute.manager [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 622.621841] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293177, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.845286] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293180, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.864111] env[62814]: DEBUG nova.compute.manager [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 623.082174] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 623.090846] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f791f94f-0ea0-4b3d-b392-65ee154963d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.099822] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c2b543-6c5f-4595-aafc-83d17620dbdf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.122020] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293177, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.706284} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.141311] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Copied Virtual Disk [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22/tmp-sparse.vmdk to [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 623.141513] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Deleting the datastore file [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22/tmp-sparse.vmdk {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 623.142081] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e969c6dd-dd67-424e-9393-ec23defd1f76 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.144440] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a70b9523-257d-482f-82d6-2eb983f422bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.157319] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65fcdba0-4259-4076-bb9e-35a81217e1c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.161393] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 623.161393] env[62814]: value = "task-4293181" [ 623.161393] env[62814]: _type = "Task" [ 623.161393] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.168688] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "62bc755d-4f96-4486-884b-0d0c337267aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 623.168928] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "62bc755d-4f96-4486-884b-0d0c337267aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 623.178813] env[62814]: DEBUG nova.compute.provider_tree [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.189619] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293181, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025613} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.189871] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 623.190090] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Moving file from [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080/e7d68632-de75-4206-8f75-4abd879e1d22 to [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22. {{(pid=62814) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 623.190350] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-e93b964e-5bdb-4829-bbfd-ad5e1c64006b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.199489] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 623.199489] env[62814]: value = "task-4293182" [ 623.199489] env[62814]: _type = "Task" [ 623.199489] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.207296] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293182, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.280367] env[62814]: DEBUG nova.compute.manager [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Received event network-changed-a896bdd8-35a5-4767-a971-88307c117164 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 623.280367] env[62814]: DEBUG nova.compute.manager [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Refreshing instance network info cache due to event network-changed-a896bdd8-35a5-4767-a971-88307c117164. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 623.280512] env[62814]: DEBUG oslo_concurrency.lockutils [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] Acquiring lock "refresh_cache-32f1bd48-3a2e-42fe-8033-85d3ad55affe" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.280648] env[62814]: DEBUG oslo_concurrency.lockutils [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] Acquired lock "refresh_cache-32f1bd48-3a2e-42fe-8033-85d3ad55affe" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.281170] env[62814]: DEBUG nova.network.neutron [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Refreshing network info cache for port a896bdd8-35a5-4767-a971-88307c117164 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 623.346060] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293180, 'name': CreateVM_Task, 'duration_secs': 0.541469} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.346617] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 623.347133] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.347568] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.347568] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 623.348012] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df2f42a0-5fd9-4dd4-bd75-f53affd8c9b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.352567] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for the task: (returnval){ [ 623.352567] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521faf8b-eb00-51b1-aa3d-77b995e307d8" [ 623.352567] env[62814]: _type = "Task" [ 623.352567] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.360683] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521faf8b-eb00-51b1-aa3d-77b995e307d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.681778] env[62814]: DEBUG nova.compute.manager [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 623.687241] env[62814]: DEBUG nova.scheduler.client.report [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 623.710512] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293182, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024949} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.711334] env[62814]: DEBUG nova.network.neutron [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Successfully updated port: 320f12a1-7803-4527-847e-8660c500ed93 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 623.712495] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] File moved {{(pid=62814) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 623.712733] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Cleaning up location [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 623.712916] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Deleting the datastore file [datastore2] vmware_temp/b47e1bff-302e-4e59-8c4a-4834ff7f7080 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 623.713484] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2dedcdf-8374-4a4c-88c1-c8b93eef0b24 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.719962] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 623.719962] env[62814]: value = "task-4293183" [ 623.719962] env[62814]: _type = "Task" [ 623.719962] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.729672] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293183, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.837063] env[62814]: DEBUG nova.network.neutron [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Successfully created port: 2710b5a0-6082-4d38-b71c-312e2c456a04 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.869822] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521faf8b-eb00-51b1-aa3d-77b995e307d8, 'name': SearchDatastore_Task, 'duration_secs': 0.009224} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.874019] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 623.874019] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 623.874019] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.874019] env[62814]: DEBUG nova.compute.manager [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 623.915838] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 623.916140] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 623.916289] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 623.916512] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 623.916640] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 623.917570] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 623.918314] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 623.918314] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 623.918314] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 623.918457] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 623.918805] env[62814]: DEBUG nova.virt.hardware [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 623.919679] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943bfec6-c803-4c07-a95a-5fbc57154f00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.932613] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53d046e-75ff-4c20-bc70-1e586b668124 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.198433] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 624.199348] env[62814]: DEBUG nova.compute.manager [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 624.205701] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.033s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 624.207259] env[62814]: INFO nova.compute.claims [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.219410] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquiring lock "refresh_cache-66db614c-d101-409d-a5c3-9de38e5a6eff" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.219410] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquired lock "refresh_cache-66db614c-d101-409d-a5c3-9de38e5a6eff" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.219564] env[62814]: DEBUG nova.network.neutron [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 624.233865] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293183, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025483} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.235457] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 624.236280] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-803971cc-673c-4e6a-a631-4ba90bd745ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.243846] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 624.243846] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5289eaa1-1e64-f5bf-ace5-e6b176f3a0bf" [ 624.243846] env[62814]: _type = "Task" [ 624.243846] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.257938] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5289eaa1-1e64-f5bf-ace5-e6b176f3a0bf, 'name': SearchDatastore_Task, 'duration_secs': 0.009967} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.257938] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.257938] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 591e4a92-7c03-4d70-885f-49a66032c97e/591e4a92-7c03-4d70-885f-49a66032c97e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 624.257938] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.258201] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 624.258201] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3cbc830b-e765-4b85-9a22-2444ec9cf861 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.261771] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2012a052-a783-42fc-a033-fb3993619179 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.272034] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 624.272034] env[62814]: value = "task-4293184" [ 624.272034] env[62814]: _type = "Task" [ 624.272034] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.275931] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 624.275931] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 624.278933] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9973ab0-ba0f-4b07-b511-98062680c308 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.289247] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293184, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.289957] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for the task: (returnval){ [ 624.289957] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526ddb40-00db-1abd-4f02-dddce5ea13fb" [ 624.289957] env[62814]: _type = "Task" [ 624.289957] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.301352] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526ddb40-00db-1abd-4f02-dddce5ea13fb, 'name': SearchDatastore_Task, 'duration_secs': 0.00902} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.304573] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e74beb4b-6e9a-4f99-9fee-5207155e77ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.311276] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for the task: (returnval){ [ 624.311276] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5290a867-2e3f-e263-bc08-182fe1042cf3" [ 624.311276] env[62814]: _type = "Task" [ 624.311276] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.315620] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 624.322356] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5290a867-2e3f-e263-bc08-182fe1042cf3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.658590] env[62814]: DEBUG nova.network.neutron [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Updated VIF entry in instance network info cache for port a896bdd8-35a5-4767-a971-88307c117164. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 624.662116] env[62814]: DEBUG nova.network.neutron [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Updating instance_info_cache with network_info: [{"id": "a896bdd8-35a5-4767-a971-88307c117164", "address": "fa:16:3e:75:f5:8f", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.161", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa896bdd8-35", "ovs_interfaceid": "a896bdd8-35a5-4767-a971-88307c117164", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.713123] env[62814]: DEBUG nova.compute.utils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 624.717723] env[62814]: DEBUG nova.compute.manager [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 624.717723] env[62814]: DEBUG nova.network.neutron [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 624.791323] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293184, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514551} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.791323] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 591e4a92-7c03-4d70-885f-49a66032c97e/591e4a92-7c03-4d70-885f-49a66032c97e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 624.791323] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 624.791323] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8134d6f9-675d-4fc9-91bd-3d5404cb5d2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.802277] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 624.802277] env[62814]: value = "task-4293185" [ 624.802277] env[62814]: _type = "Task" [ 624.802277] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.811916] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293185, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.822779] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5290a867-2e3f-e263-bc08-182fe1042cf3, 'name': SearchDatastore_Task, 'duration_secs': 0.009706} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.823094] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.823363] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 32f1bd48-3a2e-42fe-8033-85d3ad55affe/32f1bd48-3a2e-42fe-8033-85d3ad55affe.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 624.824119] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.824764] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 624.825073] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce9a0f27-776f-4003-a16d-2e7784237a33 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.827547] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b05b9f2-2909-4900-8af5-176333d59937 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.834750] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for the task: (returnval){ [ 624.834750] env[62814]: value = "task-4293186" [ 624.834750] env[62814]: _type = "Task" [ 624.834750] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.840531] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 624.840531] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 624.841269] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42a4122d-1b81-4cec-9191-7b75d35b6a11 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.849077] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293186, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.852302] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for the task: (returnval){ [ 624.852302] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5249d79e-a840-d1e0-6ca0-14bf49f04233" [ 624.852302] env[62814]: _type = "Task" [ 624.852302] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.863140] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5249d79e-a840-d1e0-6ca0-14bf49f04233, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.064420] env[62814]: DEBUG nova.network.neutron [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.165034] env[62814]: DEBUG oslo_concurrency.lockutils [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] Releasing lock "refresh_cache-32f1bd48-3a2e-42fe-8033-85d3ad55affe" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 625.165034] env[62814]: DEBUG nova.compute.manager [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Received event network-vif-plugged-5dc7f23e-f7a5-49d3-8451-88a7f80d7c90 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 625.165222] env[62814]: DEBUG oslo_concurrency.lockutils [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] Acquiring lock "37649169-66e6-4ef1-a4cd-c65b9831004a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 625.165349] env[62814]: DEBUG oslo_concurrency.lockutils [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] Lock "37649169-66e6-4ef1-a4cd-c65b9831004a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 625.165517] env[62814]: DEBUG oslo_concurrency.lockutils [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] Lock "37649169-66e6-4ef1-a4cd-c65b9831004a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.165730] env[62814]: DEBUG nova.compute.manager [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] No waiting events found dispatching network-vif-plugged-5dc7f23e-f7a5-49d3-8451-88a7f80d7c90 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 625.165852] env[62814]: WARNING nova.compute.manager [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Received unexpected event network-vif-plugged-5dc7f23e-f7a5-49d3-8451-88a7f80d7c90 for instance with vm_state building and task_state spawning. [ 625.166017] env[62814]: DEBUG nova.compute.manager [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Received event network-changed-5dc7f23e-f7a5-49d3-8451-88a7f80d7c90 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 625.166192] env[62814]: DEBUG nova.compute.manager [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Refreshing instance network info cache due to event network-changed-5dc7f23e-f7a5-49d3-8451-88a7f80d7c90. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 625.166377] env[62814]: DEBUG oslo_concurrency.lockutils [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] Acquiring lock "refresh_cache-37649169-66e6-4ef1-a4cd-c65b9831004a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.166515] env[62814]: DEBUG oslo_concurrency.lockutils [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] Acquired lock "refresh_cache-37649169-66e6-4ef1-a4cd-c65b9831004a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 625.166667] env[62814]: DEBUG nova.network.neutron [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Refreshing network info cache for port 5dc7f23e-f7a5-49d3-8451-88a7f80d7c90 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.181253] env[62814]: DEBUG nova.policy [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8606b58ab5e4b8fb00c2567d5503aa7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f72549bcb9714d3e8b1a628d982b17a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 625.217298] env[62814]: DEBUG nova.compute.manager [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 625.322870] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293185, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073765} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.323909] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 625.325404] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d52246d-7bc4-484a-a0ea-9a32f7a0595c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.357559] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 591e4a92-7c03-4d70-885f-49a66032c97e/591e4a92-7c03-4d70-885f-49a66032c97e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 625.383870] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1e29ca6e-6164-420b-9285-59631129f39d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.408183] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 625.408183] env[62814]: value = "task-4293187" [ 625.408183] env[62814]: _type = "Task" [ 625.408183] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.408183] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293186, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453657} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.409919] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 32f1bd48-3a2e-42fe-8033-85d3ad55affe/32f1bd48-3a2e-42fe-8033-85d3ad55affe.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 625.413880] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 625.417560] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-edfed955-f749-41da-83b3-f3f87a505d3c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.431994] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for the task: (returnval){ [ 625.431994] env[62814]: value = "task-4293188" [ 625.431994] env[62814]: _type = "Task" [ 625.431994] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.438341] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5249d79e-a840-d1e0-6ca0-14bf49f04233, 'name': SearchDatastore_Task, 'duration_secs': 0.008396} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.438588] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293187, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.450710] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-231e975b-a38d-4177-bdc9-f689da1e5cd8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.467942] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for the task: (returnval){ [ 625.467942] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d02e4f-4d33-d19d-cd16-17f01a847ef1" [ 625.467942] env[62814]: _type = "Task" [ 625.467942] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.468172] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293188, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.484676] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d02e4f-4d33-d19d-cd16-17f01a847ef1, 'name': SearchDatastore_Task, 'duration_secs': 0.010554} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.489828] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 625.489828] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 37649169-66e6-4ef1-a4cd-c65b9831004a/37649169-66e6-4ef1-a4cd-c65b9831004a.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 625.489985] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7803bd3a-2c40-4784-946e-330a83626896 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.506902] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for the task: (returnval){ [ 625.506902] env[62814]: value = "task-4293189" [ 625.506902] env[62814]: _type = "Task" [ 625.506902] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.517842] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293189, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.611037] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd22ce1-8120-4dc1-98f5-25132f7cbe37 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.620287] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fea309-1518-4bad-b397-86663f84d0d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.630309] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "5087b202-9bba-4489-823b-5d93cbf116e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 625.630309] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "5087b202-9bba-4489-823b-5d93cbf116e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 625.668632] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-999cdbd9-fe86-4ce3-a027-6c67b4ba6afa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.684185] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2190601b-ded1-44c7-9357-9d010d347e96 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.700688] env[62814]: DEBUG nova.compute.provider_tree [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 625.927480] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293187, 'name': ReconfigVM_Task, 'duration_secs': 0.459118} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.928792] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 591e4a92-7c03-4d70-885f-49a66032c97e/591e4a92-7c03-4d70-885f-49a66032c97e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 625.928792] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-165e16f3-dc08-4b24-af2d-25abe51b24e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.936120] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 625.936120] env[62814]: value = "task-4293190" [ 625.936120] env[62814]: _type = "Task" [ 625.936120] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.955115] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293188, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073407} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.960610] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 625.961332] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293190, 'name': Rename_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.962680] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e562854-bb5b-42ce-9252-4304f74ae2b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.993203] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] 32f1bd48-3a2e-42fe-8033-85d3ad55affe/32f1bd48-3a2e-42fe-8033-85d3ad55affe.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 625.995147] env[62814]: DEBUG nova.network.neutron [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Updating instance_info_cache with network_info: [{"id": "320f12a1-7803-4527-847e-8660c500ed93", "address": "fa:16:3e:d0:ac:4c", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.22", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap320f12a1-78", "ovs_interfaceid": "320f12a1-7803-4527-847e-8660c500ed93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.995706] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a0952e1-74f7-4252-a0a4-96269d8ebb01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.028650] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293189, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448517} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.028650] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for the task: (returnval){ [ 626.028650] env[62814]: value = "task-4293191" [ 626.028650] env[62814]: _type = "Task" [ 626.028650] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.030715] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 37649169-66e6-4ef1-a4cd-c65b9831004a/37649169-66e6-4ef1-a4cd-c65b9831004a.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 626.030946] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 626.031300] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39981881-f2c8-4f41-aac7-87320e0202bb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.047629] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for the task: (returnval){ [ 626.047629] env[62814]: value = "task-4293192" [ 626.047629] env[62814]: _type = "Task" [ 626.047629] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.060929] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293192, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.237563] env[62814]: ERROR nova.scheduler.client.report [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [req-c7c7c95a-3b0e-40ed-9c23-8bbc499139fb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c7c7c95a-3b0e-40ed-9c23-8bbc499139fb"}]} [ 626.241370] env[62814]: DEBUG nova.compute.manager [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 626.266366] env[62814]: DEBUG nova.scheduler.client.report [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 626.285972] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 626.286266] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.287420] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 626.287420] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.287420] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 626.287420] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 626.287420] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 626.287662] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 626.287662] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 626.288301] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 626.288301] env[62814]: DEBUG nova.virt.hardware [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 626.289154] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf35ab4a-21d9-42e9-b3c1-9ca01fcdea22 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.294037] env[62814]: DEBUG nova.scheduler.client.report [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 626.294037] env[62814]: DEBUG nova.compute.provider_tree [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 146, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 626.300258] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a52dfb6-84ce-42b6-8674-24d4b2c81e5f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.317184] env[62814]: DEBUG nova.scheduler.client.report [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 626.341783] env[62814]: DEBUG nova.scheduler.client.report [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 626.455164] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293190, 'name': Rename_Task, 'duration_secs': 0.158602} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.455164] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 626.455447] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ab8e01f-9b11-4b28-882e-ca7b24d0dabc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.465540] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 626.465540] env[62814]: value = "task-4293193" [ 626.465540] env[62814]: _type = "Task" [ 626.465540] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.473708] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.513559] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Releasing lock "refresh_cache-66db614c-d101-409d-a5c3-9de38e5a6eff" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 626.514278] env[62814]: DEBUG nova.compute.manager [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Instance network_info: |[{"id": "320f12a1-7803-4527-847e-8660c500ed93", "address": "fa:16:3e:d0:ac:4c", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.22", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap320f12a1-78", "ovs_interfaceid": "320f12a1-7803-4527-847e-8660c500ed93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 626.514870] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:ac:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '320f12a1-7803-4527-847e-8660c500ed93', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 626.527507] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Creating folder: Project (5f8144abf6b549cb8246aa63420e625a). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 626.529459] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f648eab-9d68-4574-8a63-01cf45f2502d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.534527] env[62814]: DEBUG nova.compute.manager [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Received event network-vif-plugged-320f12a1-7803-4527-847e-8660c500ed93 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 626.535039] env[62814]: DEBUG oslo_concurrency.lockutils [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] Acquiring lock "66db614c-d101-409d-a5c3-9de38e5a6eff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 626.535481] env[62814]: DEBUG oslo_concurrency.lockutils [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] Lock "66db614c-d101-409d-a5c3-9de38e5a6eff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 626.535840] env[62814]: DEBUG oslo_concurrency.lockutils [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] Lock "66db614c-d101-409d-a5c3-9de38e5a6eff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 626.537016] env[62814]: DEBUG nova.compute.manager [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] No waiting events found dispatching network-vif-plugged-320f12a1-7803-4527-847e-8660c500ed93 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 626.537016] env[62814]: WARNING nova.compute.manager [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Received unexpected event network-vif-plugged-320f12a1-7803-4527-847e-8660c500ed93 for instance with vm_state building and task_state spawning. [ 626.537016] env[62814]: DEBUG nova.compute.manager [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Received event network-changed-320f12a1-7803-4527-847e-8660c500ed93 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 626.537016] env[62814]: DEBUG nova.compute.manager [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Refreshing instance network info cache due to event network-changed-320f12a1-7803-4527-847e-8660c500ed93. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 626.537016] env[62814]: DEBUG oslo_concurrency.lockutils [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] Acquiring lock "refresh_cache-66db614c-d101-409d-a5c3-9de38e5a6eff" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.537574] env[62814]: DEBUG oslo_concurrency.lockutils [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] Acquired lock "refresh_cache-66db614c-d101-409d-a5c3-9de38e5a6eff" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 626.537856] env[62814]: DEBUG nova.network.neutron [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Refreshing network info cache for port 320f12a1-7803-4527-847e-8660c500ed93 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 626.561461] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293191, 'name': ReconfigVM_Task, 'duration_secs': 0.293986} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.569857] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Reconfigured VM instance instance-00000001 to attach disk [datastore2] 32f1bd48-3a2e-42fe-8033-85d3ad55affe/32f1bd48-3a2e-42fe-8033-85d3ad55affe.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 626.571399] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293192, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072987} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.571621] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Created folder: Project (5f8144abf6b549cb8246aa63420e625a) in parent group-v845547. [ 626.575037] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Creating folder: Instances. Parent ref: group-v845557. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 626.575037] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ec99623-1e29-4821-b059-59b23daaf462 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.577817] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 626.578132] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c83d24b-e3e1-4b14-9d06-55757e070c2a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.581338] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36df0834-a300-4562-8ce2-a132c652563d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.609463] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for the task: (returnval){ [ 626.609463] env[62814]: value = "task-4293195" [ 626.609463] env[62814]: _type = "Task" [ 626.609463] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.620236] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 37649169-66e6-4ef1-a4cd-c65b9831004a/37649169-66e6-4ef1-a4cd-c65b9831004a.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 626.623807] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f4f3fce-a93e-448c-9f24-d3d54176b7f3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.640290] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Created folder: Instances in parent group-v845557. [ 626.640626] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 626.646387] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 626.650140] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cdf6d757-cd32-4763-af3c-cd335d5e8cad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.673918] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for the task: (returnval){ [ 626.673918] env[62814]: value = "task-4293197" [ 626.673918] env[62814]: _type = "Task" [ 626.673918] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.676417] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7884c076-a13e-40ba-95ff-7845007c8705 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.689846] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293195, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.689846] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 626.689846] env[62814]: value = "task-4293198" [ 626.689846] env[62814]: _type = "Task" [ 626.689846] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.698215] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293197, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.700140] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac34a7e-ec55-48d2-a62c-3ad5ca97c290 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.706713] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293198, 'name': CreateVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.736396] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04560ac8-c514-4ca0-ba0d-f3bf0c380719 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.744691] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab60053-6959-44b6-9c5f-b8cabf27e9d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.764150] env[62814]: DEBUG nova.compute.provider_tree [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 626.874568] env[62814]: DEBUG nova.network.neutron [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Successfully created port: e4cf948d-498b-4313-851a-72093b0a656c {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.902909] env[62814]: DEBUG nova.network.neutron [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Updated VIF entry in instance network info cache for port 5dc7f23e-f7a5-49d3-8451-88a7f80d7c90. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 626.903207] env[62814]: DEBUG nova.network.neutron [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Updating instance_info_cache with network_info: [{"id": "5dc7f23e-f7a5-49d3-8451-88a7f80d7c90", "address": "fa:16:3e:cc:4b:b6", "network": {"id": "fe8eed21-168a-4072-b84a-df0ba6420f17", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1129990206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a5b7b2f93434cbe9df49ccc5374928c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "22b8c642-38ad-4c11-9051-145ab3bc54f2", "external-id": "nsx-vlan-transportzone-247", "segmentation_id": 247, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5dc7f23e-f7", "ovs_interfaceid": "5dc7f23e-f7a5-49d3-8451-88a7f80d7c90", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.975656] env[62814]: DEBUG oslo_vmware.api [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293193, 'name': PowerOnVM_Task, 'duration_secs': 0.473019} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.975978] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 626.976426] env[62814]: INFO nova.compute.manager [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Took 7.73 seconds to spawn the instance on the hypervisor. [ 626.976694] env[62814]: DEBUG nova.compute.manager [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 626.977623] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a105d5e5-4f1d-43c1-a08e-f012af5885fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.137169] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293195, 'name': Rename_Task, 'duration_secs': 0.166887} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.137169] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 627.137169] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9285e3eb-aceb-4654-9597-ccc64077c785 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.143834] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for the task: (returnval){ [ 627.143834] env[62814]: value = "task-4293199" [ 627.143834] env[62814]: _type = "Task" [ 627.143834] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.150755] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293199, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.188596] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293197, 'name': ReconfigVM_Task, 'duration_secs': 0.36737} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.188596] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 37649169-66e6-4ef1-a4cd-c65b9831004a/37649169-66e6-4ef1-a4cd-c65b9831004a.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 627.189350] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2475db05-fcfb-4dde-80a7-2f486e536bfd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.201207] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293198, 'name': CreateVM_Task, 'duration_secs': 0.335002} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.202935] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 627.203484] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for the task: (returnval){ [ 627.203484] env[62814]: value = "task-4293200" [ 627.203484] env[62814]: _type = "Task" [ 627.203484] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.204213] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.204324] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 627.204632] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 627.204921] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b40482f3-ef99-4fda-a4dc-ab4f9470c177 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.213061] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for the task: (returnval){ [ 627.213061] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c2b3e9-bdc1-874a-0137-093f8f168256" [ 627.213061] env[62814]: _type = "Task" [ 627.213061] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.219130] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293200, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.229089] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c2b3e9-bdc1-874a-0137-093f8f168256, 'name': SearchDatastore_Task, 'duration_secs': 0.009118} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.229397] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 627.229675] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 627.229836] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.229975] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 627.230166] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 627.230444] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ff7742b-147b-4b2f-8dec-afa72f951bf1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.238476] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 627.238705] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 627.239585] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aeb7f3ca-a48d-45ed-80be-68c1c1f74d21 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.245935] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for the task: (returnval){ [ 627.245935] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52470081-2117-20c8-829e-dd93e1c83716" [ 627.245935] env[62814]: _type = "Task" [ 627.245935] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.253342] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52470081-2117-20c8-829e-dd93e1c83716, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.323301] env[62814]: DEBUG nova.scheduler.client.report [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 15 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 627.323588] env[62814]: DEBUG nova.compute.provider_tree [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 15 to 16 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 627.323769] env[62814]: DEBUG nova.compute.provider_tree [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 627.407634] env[62814]: DEBUG oslo_concurrency.lockutils [req-77052468-4876-4c8d-b570-c5eb08eccb69 req-89b1eabb-89ad-48e4-856c-a21bcbf09806 service nova] Releasing lock "refresh_cache-37649169-66e6-4ef1-a4cd-c65b9831004a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 627.499580] env[62814]: INFO nova.compute.manager [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Took 15.33 seconds to build instance. [ 627.512546] env[62814]: DEBUG nova.network.neutron [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Successfully updated port: 2710b5a0-6082-4d38-b71c-312e2c456a04 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 627.655053] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293199, 'name': PowerOnVM_Task} progress is 37%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.717367] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293200, 'name': Rename_Task, 'duration_secs': 0.150021} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.719609] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 627.719609] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c365dbb-f8fc-4fdf-bf4a-46c9f88c6a60 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.726994] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for the task: (returnval){ [ 627.726994] env[62814]: value = "task-4293201" [ 627.726994] env[62814]: _type = "Task" [ 627.726994] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.736380] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293201, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.757685] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52470081-2117-20c8-829e-dd93e1c83716, 'name': SearchDatastore_Task, 'duration_secs': 0.008938} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.758896] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e9fcc8e-3edb-4323-af40-aaf7e2bc18a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.764495] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for the task: (returnval){ [ 627.764495] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f4130f-65d5-02c5-0607-0751180cf135" [ 627.764495] env[62814]: _type = "Task" [ 627.764495] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.776029] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f4130f-65d5-02c5-0607-0751180cf135, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.833034] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.625s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 627.833034] env[62814]: DEBUG nova.compute.manager [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 627.835901] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.901s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 627.838044] env[62814]: INFO nova.compute.claims [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.999161] env[62814]: DEBUG nova.network.neutron [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Updated VIF entry in instance network info cache for port 320f12a1-7803-4527-847e-8660c500ed93. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 627.999369] env[62814]: DEBUG nova.network.neutron [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Updating instance_info_cache with network_info: [{"id": "320f12a1-7803-4527-847e-8660c500ed93", "address": "fa:16:3e:d0:ac:4c", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.22", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap320f12a1-78", "ovs_interfaceid": "320f12a1-7803-4527-847e-8660c500ed93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.004908] env[62814]: DEBUG oslo_concurrency.lockutils [None req-031ae7f1-d641-4173-9645-f98f02dccbd7 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lock "591e4a92-7c03-4d70-885f-49a66032c97e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.846s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 628.019720] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "refresh_cache-be02bc93-1278-4b3d-afa3-270c84585d1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.019849] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquired lock "refresh_cache-be02bc93-1278-4b3d-afa3-270c84585d1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.020032] env[62814]: DEBUG nova.network.neutron [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.158576] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293199, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.240890] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293201, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.276079] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f4130f-65d5-02c5-0607-0751180cf135, 'name': SearchDatastore_Task, 'duration_secs': 0.011344} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.276440] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.277024] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 66db614c-d101-409d-a5c3-9de38e5a6eff/66db614c-d101-409d-a5c3-9de38e5a6eff.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 628.277024] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b1949180-f036-48f4-8c58-db4d28ba2de3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.283950] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for the task: (returnval){ [ 628.283950] env[62814]: value = "task-4293202" [ 628.283950] env[62814]: _type = "Task" [ 628.283950] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.295912] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293202, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.343843] env[62814]: DEBUG nova.compute.utils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 628.348588] env[62814]: DEBUG nova.compute.manager [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 628.348998] env[62814]: DEBUG nova.network.neutron [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 628.503485] env[62814]: DEBUG oslo_concurrency.lockutils [req-1ba2b668-1c33-4832-b12a-a5cb6db3edfb req-9bf78a05-993c-42f5-ae9d-e04ec9348b1a service nova] Releasing lock "refresh_cache-66db614c-d101-409d-a5c3-9de38e5a6eff" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.503726] env[62814]: DEBUG nova.compute.manager [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 628.592883] env[62814]: DEBUG nova.network.neutron [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.666611] env[62814]: DEBUG oslo_vmware.api [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293199, 'name': PowerOnVM_Task, 'duration_secs': 1.143538} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.666611] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 628.666611] env[62814]: INFO nova.compute.manager [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Took 13.99 seconds to spawn the instance on the hypervisor. [ 628.666611] env[62814]: DEBUG nova.compute.manager [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 628.668984] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0605f65a-cfe8-44bf-b5e7-da40acb62c3e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.697127] env[62814]: DEBUG nova.policy [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '275a668099f546aa9e4f1e9cc165cbfe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a4f27dd44c34eb5814db7e462dd2373', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 628.744654] env[62814]: DEBUG oslo_vmware.api [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293201, 'name': PowerOnVM_Task, 'duration_secs': 0.985268} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.744654] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 628.744654] env[62814]: INFO nova.compute.manager [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Took 11.76 seconds to spawn the instance on the hypervisor. [ 628.744654] env[62814]: DEBUG nova.compute.manager [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 628.744654] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa9058b4-18cc-4826-a1f9-ab6402c734d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.795708] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293202, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.848789] env[62814]: DEBUG nova.compute.manager [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 628.993325] env[62814]: DEBUG nova.network.neutron [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Updating instance_info_cache with network_info: [{"id": "2710b5a0-6082-4d38-b71c-312e2c456a04", "address": "fa:16:3e:c4:b3:04", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2710b5a0-60", "ovs_interfaceid": "2710b5a0-6082-4d38-b71c-312e2c456a04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.039389] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 629.124933] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4f90d5-b65c-4db9-88ed-0c966599a6e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.137227] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c25c759-bd8c-4551-a59d-2c983f27d2df {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.168121] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c697d45-e705-4e6c-bf6f-7f81a8a3a3cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.176029] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c9262f-2399-4bb1-8382-0894a0c5e266 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.193367] env[62814]: DEBUG nova.compute.provider_tree [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.196497] env[62814]: INFO nova.compute.manager [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Took 18.83 seconds to build instance. [ 629.272947] env[62814]: INFO nova.compute.manager [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Took 18.18 seconds to build instance. [ 629.305485] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293202, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.623067} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.305485] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 66db614c-d101-409d-a5c3-9de38e5a6eff/66db614c-d101-409d-a5c3-9de38e5a6eff.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 629.305485] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 629.305737] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2506916f-f01b-42b9-93cf-80314cb65a9f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.314639] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for the task: (returnval){ [ 629.314639] env[62814]: value = "task-4293203" [ 629.314639] env[62814]: _type = "Task" [ 629.314639] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.325795] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293203, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.500162] env[62814]: DEBUG nova.network.neutron [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Successfully created port: cdb5c7dc-7e71-4338-af80-0211db859c91 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 629.502234] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Releasing lock "refresh_cache-be02bc93-1278-4b3d-afa3-270c84585d1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.502515] env[62814]: DEBUG nova.compute.manager [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Instance network_info: |[{"id": "2710b5a0-6082-4d38-b71c-312e2c456a04", "address": "fa:16:3e:c4:b3:04", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2710b5a0-60", "ovs_interfaceid": "2710b5a0-6082-4d38-b71c-312e2c456a04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 629.502920] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:b3:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2710b5a0-6082-4d38-b71c-312e2c456a04', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 629.512400] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Creating folder: Project (97d11f49b6ad4c82956efa254d8995ca). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 629.513163] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a65b0cd1-f08a-47e9-81c8-404d4eb7edc2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.523762] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Created folder: Project (97d11f49b6ad4c82956efa254d8995ca) in parent group-v845547. [ 629.523960] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Creating folder: Instances. Parent ref: group-v845560. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 629.524236] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3020330-aab0-4b8d-9d61-9ad74dec2fde {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.535747] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Created folder: Instances in parent group-v845560. [ 629.536018] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 629.536687] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 629.536687] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fddafdb6-6d9c-4a77-8c40-73cf761b639b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.558374] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 629.558374] env[62814]: value = "task-4293206" [ 629.558374] env[62814]: _type = "Task" [ 629.558374] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.567131] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293206, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.586814] env[62814]: DEBUG nova.network.neutron [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Successfully updated port: e4cf948d-498b-4313-851a-72093b0a656c {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 629.697915] env[62814]: DEBUG nova.scheduler.client.report [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 629.705695] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b4e239e-c28c-4978-872e-77ae35acc406 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.351s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 629.775841] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a550988-4578-4c72-ad5a-8869bd10967b tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lock "37649169-66e6-4ef1-a4cd-c65b9831004a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.691s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 629.826077] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293203, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.129556} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.832911] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 629.832911] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e577baa-740b-4002-ae64-28ab86dcb856 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.856503] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] 66db614c-d101-409d-a5c3-9de38e5a6eff/66db614c-d101-409d-a5c3-9de38e5a6eff.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 629.856884] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d35456f9-2f95-412b-a8aa-859e90056669 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.872419] env[62814]: DEBUG nova.compute.manager [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 629.880513] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for the task: (returnval){ [ 629.880513] env[62814]: value = "task-4293207" [ 629.880513] env[62814]: _type = "Task" [ 629.880513] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.888957] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293207, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.905253] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 629.906018] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.906018] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 629.906143] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.906219] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 629.906365] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 629.906577] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 629.906781] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 629.906974] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 629.907159] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 629.907316] env[62814]: DEBUG nova.virt.hardware [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 629.908259] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e31554-6271-41f5-b9ed-f561c77408ce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.917186] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a3808b-fcd8-4bd6-adf4-9f115bba94ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.069420] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293206, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.090226] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquiring lock "refresh_cache-f5504590-823d-4c37-8226-26684f7957a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.090374] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquired lock "refresh_cache-f5504590-823d-4c37-8226-26684f7957a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.090531] env[62814]: DEBUG nova.network.neutron [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 630.209654] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 630.210237] env[62814]: DEBUG nova.compute.manager [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 630.214833] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 11.194s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 630.214833] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 630.214833] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 630.215026] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.135s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 630.216544] env[62814]: INFO nova.compute.claims [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.223709] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11160ca-d517-4b45-bbbc-5ca2944918c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.238934] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be1ab21-6586-4170-8396-c95df6db9f82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.258721] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a74e27-1557-4352-9a20-1065c45da08d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.266616] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36113b66-8318-470e-a1ff-87201cd887ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.305281] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180673MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 630.305862] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 630.396114] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293207, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.575077] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293206, 'name': CreateVM_Task, 'duration_secs': 0.748271} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.575269] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 630.576091] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.576184] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 630.576523] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 630.576990] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f17c0797-d3a7-41e7-bdfc-3fc4fc86ebf7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.584558] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 630.584558] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ca3682-5e7a-2e32-bf69-2168a6d5ab56" [ 630.584558] env[62814]: _type = "Task" [ 630.584558] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.595796] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ca3682-5e7a-2e32-bf69-2168a6d5ab56, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.691390] env[62814]: DEBUG nova.network.neutron [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.729233] env[62814]: DEBUG nova.compute.utils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 630.732560] env[62814]: DEBUG nova.compute.manager [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 630.732781] env[62814]: DEBUG nova.network.neutron [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 630.894919] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293207, 'name': ReconfigVM_Task, 'duration_secs': 0.765387} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.895263] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Reconfigured VM instance instance-00000004 to attach disk [datastore2] 66db614c-d101-409d-a5c3-9de38e5a6eff/66db614c-d101-409d-a5c3-9de38e5a6eff.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 630.895917] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6dff8909-61ac-4ef7-b7c9-160cf1efc293 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.905684] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for the task: (returnval){ [ 630.905684] env[62814]: value = "task-4293208" [ 630.905684] env[62814]: _type = "Task" [ 630.905684] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.914135] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "c7cef7f3-11db-44e1-a454-98830b465b52" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 630.914554] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "c7cef7f3-11db-44e1-a454-98830b465b52" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 630.923506] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293208, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.097163] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ca3682-5e7a-2e32-bf69-2168a6d5ab56, 'name': SearchDatastore_Task, 'duration_secs': 0.017403} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.097588] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 631.097830] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 631.098199] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.098349] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 631.098525] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 631.098902] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bda9cc41-574b-4451-ab96-e8baa50f5bda {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.109883] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 631.109883] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 631.109883] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-187352db-c21e-4ace-9872-f9bbe0d34551 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.113220] env[62814]: DEBUG nova.policy [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd147453e63fd4439829b3502a3d6db97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8003fcc64a3147a296244034cfd77c9f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 631.117572] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 631.117572] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a1899b-6280-88f4-b8ec-cb32c8a528bf" [ 631.117572] env[62814]: _type = "Task" [ 631.117572] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.126620] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a1899b-6280-88f4-b8ec-cb32c8a528bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.234679] env[62814]: DEBUG nova.compute.manager [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 631.338494] env[62814]: DEBUG nova.network.neutron [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Updating instance_info_cache with network_info: [{"id": "e4cf948d-498b-4313-851a-72093b0a656c", "address": "fa:16:3e:a4:32:24", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4cf948d-49", "ovs_interfaceid": "e4cf948d-498b-4313-851a-72093b0a656c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.418966] env[62814]: DEBUG nova.compute.manager [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 631.421647] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293208, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.554498] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78538897-ee96-43b9-b981-637e73bd0191 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.563498] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7343753c-3f1a-42c6-836a-8896d3f0d335 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.597327] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2a6562-5ecf-421f-ba36-6cf23a7a380f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.605147] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e75907-f39c-4100-b032-0c17f1e188e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.630055] env[62814]: DEBUG nova.compute.provider_tree [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.640725] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a1899b-6280-88f4-b8ec-cb32c8a528bf, 'name': SearchDatastore_Task, 'duration_secs': 0.010785} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.644027] env[62814]: DEBUG nova.network.neutron [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Successfully updated port: cdb5c7dc-7e71-4338-af80-0211db859c91 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 631.644027] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7dc7201c-a288-4132-aa21-20a1a16939b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.650147] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 631.650147] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527893ce-a509-12e8-aa14-910317092238" [ 631.650147] env[62814]: _type = "Task" [ 631.650147] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.666393] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527893ce-a509-12e8-aa14-910317092238, 'name': SearchDatastore_Task, 'duration_secs': 0.011065} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.666733] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 631.666996] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] be02bc93-1278-4b3d-afa3-270c84585d1d/be02bc93-1278-4b3d-afa3-270c84585d1d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 631.668027] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0eecd05-f3ec-4461-91b0-95b8df304b42 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.674630] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 631.674630] env[62814]: value = "task-4293209" [ 631.674630] env[62814]: _type = "Task" [ 631.674630] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.684746] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293209, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.697891] env[62814]: DEBUG nova.compute.manager [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Received event network-vif-plugged-2710b5a0-6082-4d38-b71c-312e2c456a04 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 631.698172] env[62814]: DEBUG oslo_concurrency.lockutils [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] Acquiring lock "be02bc93-1278-4b3d-afa3-270c84585d1d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 631.698390] env[62814]: DEBUG oslo_concurrency.lockutils [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] Lock "be02bc93-1278-4b3d-afa3-270c84585d1d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 631.698589] env[62814]: DEBUG oslo_concurrency.lockutils [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] Lock "be02bc93-1278-4b3d-afa3-270c84585d1d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 631.698775] env[62814]: DEBUG nova.compute.manager [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] No waiting events found dispatching network-vif-plugged-2710b5a0-6082-4d38-b71c-312e2c456a04 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 631.698970] env[62814]: WARNING nova.compute.manager [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Received unexpected event network-vif-plugged-2710b5a0-6082-4d38-b71c-312e2c456a04 for instance with vm_state building and task_state spawning. [ 631.699187] env[62814]: DEBUG nova.compute.manager [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Received event network-changed-2710b5a0-6082-4d38-b71c-312e2c456a04 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 631.699373] env[62814]: DEBUG nova.compute.manager [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Refreshing instance network info cache due to event network-changed-2710b5a0-6082-4d38-b71c-312e2c456a04. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 631.699733] env[62814]: DEBUG oslo_concurrency.lockutils [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] Acquiring lock "refresh_cache-be02bc93-1278-4b3d-afa3-270c84585d1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.699831] env[62814]: DEBUG oslo_concurrency.lockutils [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] Acquired lock "refresh_cache-be02bc93-1278-4b3d-afa3-270c84585d1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 631.699943] env[62814]: DEBUG nova.network.neutron [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Refreshing network info cache for port 2710b5a0-6082-4d38-b71c-312e2c456a04 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 631.846329] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Releasing lock "refresh_cache-f5504590-823d-4c37-8226-26684f7957a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 631.846634] env[62814]: DEBUG nova.compute.manager [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Instance network_info: |[{"id": "e4cf948d-498b-4313-851a-72093b0a656c", "address": "fa:16:3e:a4:32:24", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4cf948d-49", "ovs_interfaceid": "e4cf948d-498b-4313-851a-72093b0a656c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 631.847084] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:32:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e4cf948d-498b-4313-851a-72093b0a656c', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 631.854865] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Creating folder: Project (f72549bcb9714d3e8b1a628d982b17a6). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 631.855214] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0822ac6-407d-4fd3-9218-9eb6624d6ed8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.865695] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Created folder: Project (f72549bcb9714d3e8b1a628d982b17a6) in parent group-v845547. [ 631.866084] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Creating folder: Instances. Parent ref: group-v845563. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 631.866322] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0a0b3414-021c-47a5-98ac-6f7aece75a5d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.876866] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Created folder: Instances in parent group-v845563. [ 631.879710] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 631.879710] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5504590-823d-4c37-8226-26684f7957a6] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 631.879710] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6f0c0b4-7182-4a49-8deb-de4001d5dcc7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.898038] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 631.898038] env[62814]: value = "task-4293212" [ 631.898038] env[62814]: _type = "Task" [ 631.898038] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.906425] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293212, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.921520] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293208, 'name': Rename_Task, 'duration_secs': 0.863358} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.921872] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 631.922211] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d9c5df5-4341-4b06-a286-4fa2a54511c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.930962] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for the task: (returnval){ [ 631.930962] env[62814]: value = "task-4293213" [ 631.930962] env[62814]: _type = "Task" [ 631.930962] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.952716] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293213, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.959643] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.136428] env[62814]: DEBUG nova.scheduler.client.report [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 632.145017] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "refresh_cache-0ebb226a-40a3-40cd-8492-ccce87baf0e6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.146030] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquired lock "refresh_cache-0ebb226a-40a3-40cd-8492-ccce87baf0e6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 632.146030] env[62814]: DEBUG nova.network.neutron [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.174583] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquiring lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 632.174583] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.189560] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293209, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.248281] env[62814]: DEBUG nova.compute.manager [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 632.293672] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 632.293672] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.293672] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 632.293672] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.293975] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 632.293975] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 632.293975] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 632.293975] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 632.293975] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 632.294332] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 632.294332] env[62814]: DEBUG nova.virt.hardware [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 632.294332] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1e70c8-500e-48d3-a045-323c4eef7c3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.313235] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8564f234-12c6-4a9e-adf5-a910cdfdec82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.355344] env[62814]: DEBUG nova.compute.manager [None req-53c19005-fc04-446a-9306-0a8bc4cbfb34 tempest-ServerDiagnosticsV248Test-1633039967 tempest-ServerDiagnosticsV248Test-1633039967-project-admin] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 632.357087] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d85a59d-bdc1-497a-9bd7-05a2871e9aea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.366179] env[62814]: INFO nova.compute.manager [None req-53c19005-fc04-446a-9306-0a8bc4cbfb34 tempest-ServerDiagnosticsV248Test-1633039967 tempest-ServerDiagnosticsV248Test-1633039967-project-admin] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Retrieving diagnostics [ 632.367214] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4abee8-8e6d-43f6-a4b9-78f76942ce49 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.372264] env[62814]: DEBUG nova.network.neutron [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Successfully created port: 827a94b4-5864-4060-bfb5-d0e9d2281332 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 632.418618] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293212, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.450128] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293213, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.642658] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 632.643739] env[62814]: DEBUG nova.compute.manager [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 632.646461] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.331s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 632.647997] env[62814]: INFO nova.compute.claims [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 632.680019] env[62814]: DEBUG nova.compute.manager [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 632.694813] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293209, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582502} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.695035] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] be02bc93-1278-4b3d-afa3-270c84585d1d/be02bc93-1278-4b3d-afa3-270c84585d1d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 632.695304] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 632.695575] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-451af6b7-0f70-4eaf-ad1d-28ceb694ccfa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.703955] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 632.703955] env[62814]: value = "task-4293214" [ 632.703955] env[62814]: _type = "Task" [ 632.703955] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.714493] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293214, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.735402] env[62814]: DEBUG nova.network.neutron [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.769655] env[62814]: DEBUG nova.network.neutron [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Updated VIF entry in instance network info cache for port 2710b5a0-6082-4d38-b71c-312e2c456a04. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 632.770407] env[62814]: DEBUG nova.network.neutron [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Updating instance_info_cache with network_info: [{"id": "2710b5a0-6082-4d38-b71c-312e2c456a04", "address": "fa:16:3e:c4:b3:04", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2710b5a0-60", "ovs_interfaceid": "2710b5a0-6082-4d38-b71c-312e2c456a04", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.919020] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293212, 'name': CreateVM_Task, 'duration_secs': 0.696091} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.919302] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5504590-823d-4c37-8226-26684f7957a6] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 632.919834] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.919921] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 632.920227] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 632.920470] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa464d11-7535-4177-b204-5268c594e133 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.925240] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for the task: (returnval){ [ 632.925240] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5268f5d2-7f07-8122-c3bd-534cc3b0587e" [ 632.925240] env[62814]: _type = "Task" [ 632.925240] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.933202] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5268f5d2-7f07-8122-c3bd-534cc3b0587e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.935996] env[62814]: DEBUG nova.network.neutron [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Updating instance_info_cache with network_info: [{"id": "cdb5c7dc-7e71-4338-af80-0211db859c91", "address": "fa:16:3e:d3:01:ec", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb5c7dc-7e", "ovs_interfaceid": "cdb5c7dc-7e71-4338-af80-0211db859c91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.944899] env[62814]: DEBUG oslo_vmware.api [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293213, 'name': PowerOnVM_Task, 'duration_secs': 1.00112} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.944899] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 632.945078] env[62814]: INFO nova.compute.manager [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Took 11.38 seconds to spawn the instance on the hypervisor. [ 632.945259] env[62814]: DEBUG nova.compute.manager [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 632.946233] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815804fb-32d5-428c-9636-51613dee758e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.155729] env[62814]: DEBUG nova.compute.utils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 633.155729] env[62814]: DEBUG nova.compute.manager [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 633.155729] env[62814]: DEBUG nova.network.neutron [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 633.224311] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293214, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069111} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.224756] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.225055] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 633.225892] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce6dfca-6291-4b98-bd8d-d42aa1ab0c8c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.254006] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] be02bc93-1278-4b3d-afa3-270c84585d1d/be02bc93-1278-4b3d-afa3-270c84585d1d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 633.255753] env[62814]: DEBUG nova.policy [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b96cac63d9f47e8942a0e9e3e9c96bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd5f0190bd5f24a909c9839b7fd1e870b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 633.257842] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2bc15977-1e5c-4f58-b634-3d6174a5073f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.272603] env[62814]: DEBUG oslo_concurrency.lockutils [req-9e3a00b0-0012-4c00-a12d-2bb1023d2938 req-beebc3d2-916c-416c-a806-ccd20bfb6107 service nova] Releasing lock "refresh_cache-be02bc93-1278-4b3d-afa3-270c84585d1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.279459] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 633.279459] env[62814]: value = "task-4293215" [ 633.279459] env[62814]: _type = "Task" [ 633.279459] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.287301] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293215, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.440129] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5268f5d2-7f07-8122-c3bd-534cc3b0587e, 'name': SearchDatastore_Task, 'duration_secs': 0.010433} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.440129] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.440129] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 633.440129] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.440287] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.440287] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 633.440287] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1f903d8-8f38-4f00-9294-ef751464cd5f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.442943] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Releasing lock "refresh_cache-0ebb226a-40a3-40cd-8492-ccce87baf0e6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.442943] env[62814]: DEBUG nova.compute.manager [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Instance network_info: |[{"id": "cdb5c7dc-7e71-4338-af80-0211db859c91", "address": "fa:16:3e:d3:01:ec", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb5c7dc-7e", "ovs_interfaceid": "cdb5c7dc-7e71-4338-af80-0211db859c91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 633.443584] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d3:01:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cdb5c7dc-7e71-4338-af80-0211db859c91', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 633.454854] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Creating folder: Project (1a4f27dd44c34eb5814db7e462dd2373). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 633.456695] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8071a11c-b6d7-473d-ad72-ca9494b8319d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.467093] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 633.467460] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 633.474181] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1e8ccb8-bb33-4ab6-963e-40858f41ce99 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.474181] env[62814]: INFO nova.compute.manager [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Took 20.67 seconds to build instance. [ 633.478526] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for the task: (returnval){ [ 633.478526] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5277d2ac-b29a-f58f-292d-f3d009c6911f" [ 633.478526] env[62814]: _type = "Task" [ 633.478526] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.483604] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Created folder: Project (1a4f27dd44c34eb5814db7e462dd2373) in parent group-v845547. [ 633.484208] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Creating folder: Instances. Parent ref: group-v845566. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 633.484670] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ec10d918-8045-4f41-aea8-99157d5810b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.493305] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5277d2ac-b29a-f58f-292d-f3d009c6911f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.496392] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Created folder: Instances in parent group-v845566. [ 633.496599] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 633.496905] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 633.500017] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d95f4c42-d26d-4a87-8c82-f2440bd36f66 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.524019] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 633.524019] env[62814]: value = "task-4293218" [ 633.524019] env[62814]: _type = "Task" [ 633.524019] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.536549] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293218, 'name': CreateVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.658866] env[62814]: DEBUG nova.compute.manager [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 633.792621] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293215, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.940973] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd99de5e-3f44-4c27-866b-7568c09e50a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.949477] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c86f22-85a8-4dca-931d-9297de0f34b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.982571] env[62814]: DEBUG oslo_concurrency.lockutils [None req-005819a0-50ed-4e80-ad6d-5ff66bedc43c tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lock "66db614c-d101-409d-a5c3-9de38e5a6eff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.194s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 633.986993] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd32a2e7-84d3-448e-9b50-35806273f945 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.996291] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5277d2ac-b29a-f58f-292d-f3d009c6911f, 'name': SearchDatastore_Task, 'duration_secs': 0.015408} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.001109] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0355bd85-8a73-4fab-a6d9-ff8f7404be70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.004750] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4689038-f094-4b13-8a5d-67a7b9028f82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.011713] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for the task: (returnval){ [ 634.011713] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528aa837-d48c-2647-53bb-e6f1a83409a8" [ 634.011713] env[62814]: _type = "Task" [ 634.011713] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.020095] env[62814]: DEBUG nova.compute.provider_tree [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.028701] env[62814]: DEBUG nova.network.neutron [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Successfully created port: 6e1ae96b-a9fd-4e11-9803-191a35a77cb2 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.034426] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528aa837-d48c-2647-53bb-e6f1a83409a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.041061] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293218, 'name': CreateVM_Task, 'duration_secs': 0.453627} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.041061] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 634.041061] env[62814]: DEBUG nova.compute.manager [req-6ff4bcf0-7f61-463d-b9ee-db8d0a933db3 req-af278d08-b591-43fc-bb6e-ca146ef3363d service nova] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Received event network-vif-plugged-cdb5c7dc-7e71-4338-af80-0211db859c91 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 634.041061] env[62814]: DEBUG oslo_concurrency.lockutils [req-6ff4bcf0-7f61-463d-b9ee-db8d0a933db3 req-af278d08-b591-43fc-bb6e-ca146ef3363d service nova] Acquiring lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 634.041061] env[62814]: DEBUG oslo_concurrency.lockutils [req-6ff4bcf0-7f61-463d-b9ee-db8d0a933db3 req-af278d08-b591-43fc-bb6e-ca146ef3363d service nova] Lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 634.041636] env[62814]: DEBUG oslo_concurrency.lockutils [req-6ff4bcf0-7f61-463d-b9ee-db8d0a933db3 req-af278d08-b591-43fc-bb6e-ca146ef3363d service nova] Lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 634.041636] env[62814]: DEBUG nova.compute.manager [req-6ff4bcf0-7f61-463d-b9ee-db8d0a933db3 req-af278d08-b591-43fc-bb6e-ca146ef3363d service nova] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] No waiting events found dispatching network-vif-plugged-cdb5c7dc-7e71-4338-af80-0211db859c91 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 634.041636] env[62814]: WARNING nova.compute.manager [req-6ff4bcf0-7f61-463d-b9ee-db8d0a933db3 req-af278d08-b591-43fc-bb6e-ca146ef3363d service nova] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Received unexpected event network-vif-plugged-cdb5c7dc-7e71-4338-af80-0211db859c91 for instance with vm_state building and task_state spawning. [ 634.043907] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.044079] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 634.044392] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 634.045247] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b56b7963-81fe-4fc0-a1f1-0918c133dcba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.050245] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 634.050245] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7a2b5-e279-a1ac-d4c5-03efad91715c" [ 634.050245] env[62814]: _type = "Task" [ 634.050245] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.058151] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7a2b5-e279-a1ac-d4c5-03efad91715c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.290164] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293215, 'name': ReconfigVM_Task, 'duration_secs': 0.518853} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.290164] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Reconfigured VM instance instance-00000005 to attach disk [datastore2] be02bc93-1278-4b3d-afa3-270c84585d1d/be02bc93-1278-4b3d-afa3-270c84585d1d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 634.290599] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0a116ee9-a382-457c-b68a-f91c77bc9f11 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.296820] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 634.296820] env[62814]: value = "task-4293219" [ 634.296820] env[62814]: _type = "Task" [ 634.296820] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.308072] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293219, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.526144] env[62814]: DEBUG nova.scheduler.client.report [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 634.533950] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528aa837-d48c-2647-53bb-e6f1a83409a8, 'name': SearchDatastore_Task, 'duration_secs': 0.020592} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.534721] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.535015] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] f5504590-823d-4c37-8226-26684f7957a6/f5504590-823d-4c37-8226-26684f7957a6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 634.535279] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43dad3a8-6eaf-4327-9ef6-82fa475e1df4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.544163] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for the task: (returnval){ [ 634.544163] env[62814]: value = "task-4293220" [ 634.544163] env[62814]: _type = "Task" [ 634.544163] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.557250] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293220, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.563590] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7a2b5-e279-a1ac-d4c5-03efad91715c, 'name': SearchDatastore_Task, 'duration_secs': 0.041589} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.563923] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.564173] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 634.564457] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.564638] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 634.564822] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 634.565544] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-78e8f997-7472-4590-9d9a-3036ab3fe85a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.585748] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 634.585938] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 634.586756] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a7c1fdc-2b40-4fe7-b625-87a90b4b9e69 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.593529] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 634.593529] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5243bce1-1373-8613-9aa2-faf93178cc53" [ 634.593529] env[62814]: _type = "Task" [ 634.593529] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.604015] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5243bce1-1373-8613-9aa2-faf93178cc53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.677771] env[62814]: DEBUG nova.compute.manager [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 634.720335] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 634.723031] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 634.723031] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 634.723031] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 634.723031] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 634.723031] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 634.723731] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 634.723731] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 634.723731] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 634.723875] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 634.724069] env[62814]: DEBUG nova.virt.hardware [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 634.725073] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02101503-6ea6-433a-a8fa-d8cd35fde04b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.738087] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87793ad4-d42c-4966-99f7-842fb580c9dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.749433] env[62814]: DEBUG nova.network.neutron [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Successfully updated port: 827a94b4-5864-4060-bfb5-d0e9d2281332 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 634.812884] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293219, 'name': Rename_Task, 'duration_secs': 0.314727} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.812884] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 634.815094] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-964102a0-0c7b-46f4-9d80-c676736eda03 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.821447] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 634.821447] env[62814]: value = "task-4293221" [ 634.821447] env[62814]: _type = "Task" [ 634.821447] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.835106] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293221, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.036433] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.036654] env[62814]: DEBUG nova.compute.manager [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 635.043254] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.004s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.045422] env[62814]: INFO nova.compute.claims [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.065253] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293220, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.111022] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5243bce1-1373-8613-9aa2-faf93178cc53, 'name': SearchDatastore_Task, 'duration_secs': 0.017465} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.114123] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ac6909e-310c-4023-a07a-8db8bbf39ada {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.120662] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 635.120662] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52467ac8-99a4-b43a-96bd-d5e9f935e533" [ 635.120662] env[62814]: _type = "Task" [ 635.120662] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.135786] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52467ac8-99a4-b43a-96bd-d5e9f935e533, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.252651] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.252795] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.252944] env[62814]: DEBUG nova.network.neutron [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 635.339764] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293221, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.510698] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquiring lock "6976b964-a8d3-4886-8aac-8d513e721018" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.510969] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "6976b964-a8d3-4886-8aac-8d513e721018" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.550365] env[62814]: DEBUG nova.compute.utils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 635.551995] env[62814]: DEBUG nova.compute.manager [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 635.551995] env[62814]: DEBUG nova.network.neutron [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 635.571449] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293220, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.603346} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.572452] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] f5504590-823d-4c37-8226-26684f7957a6/f5504590-823d-4c37-8226-26684f7957a6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 635.572452] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 635.572452] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-61a434c0-c69a-4a5e-a7e8-3e3c8d78de4e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.583644] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for the task: (returnval){ [ 635.583644] env[62814]: value = "task-4293222" [ 635.583644] env[62814]: _type = "Task" [ 635.583644] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.593244] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293222, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.631102] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52467ac8-99a4-b43a-96bd-d5e9f935e533, 'name': SearchDatastore_Task, 'duration_secs': 0.014607} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.631508] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.631680] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 0ebb226a-40a3-40cd-8492-ccce87baf0e6/0ebb226a-40a3-40cd-8492-ccce87baf0e6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 635.632015] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45f7ad79-9f8b-47ee-9bc6-669891a0c622 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.639628] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 635.639628] env[62814]: value = "task-4293223" [ 635.639628] env[62814]: _type = "Task" [ 635.639628] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.649282] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293223, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.819746] env[62814]: DEBUG nova.compute.manager [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] [instance: f5504590-823d-4c37-8226-26684f7957a6] Received event network-vif-plugged-e4cf948d-498b-4313-851a-72093b0a656c {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 635.819746] env[62814]: DEBUG oslo_concurrency.lockutils [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] Acquiring lock "f5504590-823d-4c37-8226-26684f7957a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.820193] env[62814]: DEBUG oslo_concurrency.lockutils [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] Lock "f5504590-823d-4c37-8226-26684f7957a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.820512] env[62814]: DEBUG oslo_concurrency.lockutils [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] Lock "f5504590-823d-4c37-8226-26684f7957a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.820849] env[62814]: DEBUG nova.compute.manager [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] [instance: f5504590-823d-4c37-8226-26684f7957a6] No waiting events found dispatching network-vif-plugged-e4cf948d-498b-4313-851a-72093b0a656c {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 635.821155] env[62814]: WARNING nova.compute.manager [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] [instance: f5504590-823d-4c37-8226-26684f7957a6] Received unexpected event network-vif-plugged-e4cf948d-498b-4313-851a-72093b0a656c for instance with vm_state building and task_state spawning. [ 635.822677] env[62814]: DEBUG nova.compute.manager [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] [instance: f5504590-823d-4c37-8226-26684f7957a6] Received event network-changed-e4cf948d-498b-4313-851a-72093b0a656c {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 635.823017] env[62814]: DEBUG nova.compute.manager [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] [instance: f5504590-823d-4c37-8226-26684f7957a6] Refreshing instance network info cache due to event network-changed-e4cf948d-498b-4313-851a-72093b0a656c. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 635.823371] env[62814]: DEBUG oslo_concurrency.lockutils [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] Acquiring lock "refresh_cache-f5504590-823d-4c37-8226-26684f7957a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.823649] env[62814]: DEBUG oslo_concurrency.lockutils [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] Acquired lock "refresh_cache-f5504590-823d-4c37-8226-26684f7957a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.825018] env[62814]: DEBUG nova.network.neutron [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] [instance: f5504590-823d-4c37-8226-26684f7957a6] Refreshing network info cache for port e4cf948d-498b-4313-851a-72093b0a656c {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 635.826051] env[62814]: DEBUG nova.network.neutron [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.846361] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293221, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.847926] env[62814]: DEBUG nova.policy [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '58359539fd4545cdbf200ec364e43834', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dedbeff832a4ac48b0aa01bd6acc3f1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 635.849469] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquiring lock "37649169-66e6-4ef1-a4cd-c65b9831004a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.849701] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lock "37649169-66e6-4ef1-a4cd-c65b9831004a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.850157] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquiring lock "37649169-66e6-4ef1-a4cd-c65b9831004a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.850157] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lock "37649169-66e6-4ef1-a4cd-c65b9831004a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.850250] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lock "37649169-66e6-4ef1-a4cd-c65b9831004a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.855618] env[62814]: INFO nova.compute.manager [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Terminating instance [ 635.978966] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquiring lock "296f6c11-7108-42e6-8ada-5d8c08b00da6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.979597] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lock "296f6c11-7108-42e6-8ada-5d8c08b00da6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 636.013925] env[62814]: DEBUG nova.compute.manager [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 636.063568] env[62814]: DEBUG nova.compute.manager [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 636.095869] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293222, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07702} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.099358] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 636.100588] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6432ea-a2a3-4919-81ef-17c0ced20fbc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.129250] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Reconfiguring VM instance instance-00000006 to attach disk [datastore2] f5504590-823d-4c37-8226-26684f7957a6/f5504590-823d-4c37-8226-26684f7957a6.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 636.138977] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eae4a910-4291-4114-af63-45bbd99baa25 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.173934] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293223, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.174964] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for the task: (returnval){ [ 636.174964] env[62814]: value = "task-4293224" [ 636.174964] env[62814]: _type = "Task" [ 636.174964] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.188760] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293224, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.272152] env[62814]: DEBUG nova.network.neutron [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance_info_cache with network_info: [{"id": "827a94b4-5864-4060-bfb5-d0e9d2281332", "address": "fa:16:3e:a9:8f:95", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.43", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap827a94b4-58", "ovs_interfaceid": "827a94b4-5864-4060-bfb5-d0e9d2281332", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.342448] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293221, 'name': PowerOnVM_Task} progress is 86%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.366558] env[62814]: DEBUG nova.compute.manager [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 636.366558] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.367230] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701b7dc9-90d6-4bdd-a0f9-05dab8b5e3ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.376719] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 636.379609] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-02d7a88e-b77f-4883-b775-3b374cdbfdbc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.387651] env[62814]: DEBUG oslo_vmware.api [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for the task: (returnval){ [ 636.387651] env[62814]: value = "task-4293225" [ 636.387651] env[62814]: _type = "Task" [ 636.387651] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.396640] env[62814]: DEBUG oslo_vmware.api [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293225, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.398481] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02456351-e34d-44e5-b8bd-d96b689f1f9b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.406861] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd2b43b-6e41-494c-a0ac-2178628f1d92 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.448343] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542f0b54-c23d-4303-b1a6-3570c72e9c22 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.459629] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2dd1d0-9ea4-4390-8c70-ed185aa95149 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.474555] env[62814]: DEBUG nova.compute.provider_tree [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.539386] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.673743] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293223, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634438} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.674126] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 0ebb226a-40a3-40cd-8492-ccce87baf0e6/0ebb226a-40a3-40cd-8492-ccce87baf0e6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 636.674394] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 636.674704] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ce11f443-aa9b-41ff-ac87-015901420509 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.688075] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 636.688075] env[62814]: value = "task-4293226" [ 636.688075] env[62814]: _type = "Task" [ 636.688075] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.695044] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293224, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.700521] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293226, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.776911] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.777510] env[62814]: DEBUG nova.compute.manager [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Instance network_info: |[{"id": "827a94b4-5864-4060-bfb5-d0e9d2281332", "address": "fa:16:3e:a9:8f:95", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.43", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap827a94b4-58", "ovs_interfaceid": "827a94b4-5864-4060-bfb5-d0e9d2281332", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 636.778603] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a9:8f:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '827a94b4-5864-4060-bfb5-d0e9d2281332', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 636.786269] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Creating folder: Project (8003fcc64a3147a296244034cfd77c9f). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 636.786424] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6c71b1c-721b-4543-a82f-6a8d7d2ff865 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.802092] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Created folder: Project (8003fcc64a3147a296244034cfd77c9f) in parent group-v845547. [ 636.802280] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Creating folder: Instances. Parent ref: group-v845569. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 636.802622] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eba6fc93-7d22-4a27-8bbb-2d5a7a45b57d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.813897] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Created folder: Instances in parent group-v845569. [ 636.814162] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 636.814361] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 636.814874] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c885f94-09ff-4f08-ba66-a1e8a85c4f7d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.837539] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 636.837539] env[62814]: value = "task-4293229" [ 636.837539] env[62814]: _type = "Task" [ 636.837539] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.840602] env[62814]: DEBUG oslo_vmware.api [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293221, 'name': PowerOnVM_Task, 'duration_secs': 1.66927} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.843774] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 636.844036] env[62814]: INFO nova.compute.manager [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Took 12.97 seconds to spawn the instance on the hypervisor. [ 636.844225] env[62814]: DEBUG nova.compute.manager [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 636.844935] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c09d7bbf-9754-4699-9c43-daf4877ec1f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.852506] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293229, 'name': CreateVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.858136] env[62814]: DEBUG nova.network.neutron [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Successfully updated port: 6e1ae96b-a9fd-4e11-9803-191a35a77cb2 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 636.903469] env[62814]: DEBUG oslo_vmware.api [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293225, 'name': PowerOffVM_Task, 'duration_secs': 0.270336} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.903469] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 636.903757] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 636.903909] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a233af2-fb2d-464a-9c1d-3d5e6f37809a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.952888] env[62814]: DEBUG nova.network.neutron [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Successfully created port: 559634c5-106d-4de8-b7b2-596fd81a0c30 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.972829] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 636.973054] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 636.973242] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Deleting the datastore file [datastore2] 37649169-66e6-4ef1-a4cd-c65b9831004a {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 636.973636] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4831a958-8fd2-4cd5-a9bf-fed254bef801 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.977926] env[62814]: DEBUG nova.scheduler.client.report [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 636.982147] env[62814]: DEBUG oslo_vmware.api [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for the task: (returnval){ [ 636.982147] env[62814]: value = "task-4293231" [ 636.982147] env[62814]: _type = "Task" [ 636.982147] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.990918] env[62814]: DEBUG oslo_vmware.api [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293231, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.064713] env[62814]: DEBUG nova.network.neutron [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] [instance: f5504590-823d-4c37-8226-26684f7957a6] Updated VIF entry in instance network info cache for port e4cf948d-498b-4313-851a-72093b0a656c. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 637.065388] env[62814]: DEBUG nova.network.neutron [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] [instance: f5504590-823d-4c37-8226-26684f7957a6] Updating instance_info_cache with network_info: [{"id": "e4cf948d-498b-4313-851a-72093b0a656c", "address": "fa:16:3e:a4:32:24", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape4cf948d-49", "ovs_interfaceid": "e4cf948d-498b-4313-851a-72093b0a656c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.083510] env[62814]: DEBUG nova.compute.manager [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 637.129278] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 637.129359] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.129761] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 637.129761] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.129894] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 637.129950] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 637.130387] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 637.130481] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 637.130599] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 637.130868] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 637.130948] env[62814]: DEBUG nova.virt.hardware [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 637.132159] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d654edc4-5fad-440f-8fb3-4a47fbd6d8a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.143565] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd35969e-3e0c-4098-b00c-42d55da83a4a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.192074] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293224, 'name': ReconfigVM_Task, 'duration_secs': 0.551377} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.195889] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Reconfigured VM instance instance-00000006 to attach disk [datastore2] f5504590-823d-4c37-8226-26684f7957a6/f5504590-823d-4c37-8226-26684f7957a6.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 637.196553] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f92294c7-fd01-475f-aca4-17ea9ff7aa68 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.204495] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293226, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088825} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.206325] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 637.206663] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for the task: (returnval){ [ 637.206663] env[62814]: value = "task-4293232" [ 637.206663] env[62814]: _type = "Task" [ 637.206663] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.207357] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c192244f-ac37-4c3e-a60d-460fbd07ec27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.220291] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293232, 'name': Rename_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.246697] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 0ebb226a-40a3-40cd-8492-ccce87baf0e6/0ebb226a-40a3-40cd-8492-ccce87baf0e6.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 637.246697] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c787c03-04ef-4958-ad6b-b34b36f4ec3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.270241] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 637.270241] env[62814]: value = "task-4293233" [ 637.270241] env[62814]: _type = "Task" [ 637.270241] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.281359] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293233, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.296965] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "d9242042-6209-4b04-bf00-00dd04d9d6a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 637.296965] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "d9242042-6209-4b04-bf00-00dd04d9d6a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.352612] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293229, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.369233] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "refresh_cache-17bfe703-ff96-4cb7-8535-eed57e10d673" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.369233] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquired lock "refresh_cache-17bfe703-ff96-4cb7-8535-eed57e10d673" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 637.369233] env[62814]: DEBUG nova.network.neutron [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.369865] env[62814]: INFO nova.compute.manager [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Took 24.13 seconds to build instance. [ 637.483908] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.441s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 637.484500] env[62814]: DEBUG nova.compute.manager [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 637.488058] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 7.182s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 637.503469] env[62814]: DEBUG oslo_vmware.api [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Task: {'id': task-4293231, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.413097} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.503821] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 637.504320] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 637.504571] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 637.504781] env[62814]: INFO nova.compute.manager [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 637.505103] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 637.505545] env[62814]: DEBUG nova.compute.manager [-] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 637.505545] env[62814]: DEBUG nova.network.neutron [-] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.568387] env[62814]: DEBUG oslo_concurrency.lockutils [req-b87a5163-0413-45e6-96df-cc83d2205b6c req-b0194300-919c-4d8f-a579-9da44275ad23 service nova] Releasing lock "refresh_cache-f5504590-823d-4c37-8226-26684f7957a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 637.731698] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293232, 'name': Rename_Task, 'duration_secs': 0.218574} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.731698] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 637.731698] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eb85c2f1-b16a-4748-8f79-23f147f786a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.741748] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for the task: (returnval){ [ 637.741748] env[62814]: value = "task-4293234" [ 637.741748] env[62814]: _type = "Task" [ 637.741748] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.750980] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293234, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.785341] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293233, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.852597] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293229, 'name': CreateVM_Task, 'duration_secs': 0.582202} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.852981] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 637.853915] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.854192] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 637.855407] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 637.855771] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55bb20c5-926c-4534-8308-77905e8470e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.863142] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 637.863142] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dbaee4-8c5b-9fbe-22e3-1c7aa2a24822" [ 637.863142] env[62814]: _type = "Task" [ 637.863142] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.872961] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dbaee4-8c5b-9fbe-22e3-1c7aa2a24822, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.873286] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ca76ff5f-3977-4c63-8b15-6165946c9044 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "be02bc93-1278-4b3d-afa3-270c84585d1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.639s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.000586] env[62814]: DEBUG nova.compute.utils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 638.010531] env[62814]: DEBUG nova.compute.manager [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 638.010531] env[62814]: DEBUG nova.network.neutron [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 638.118144] env[62814]: DEBUG nova.network.neutron [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.155057] env[62814]: DEBUG nova.policy [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '093788d5d63e4c6a84750ffaf1e31455', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d51711cd907404b8d0a48d904b599c4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 638.259257] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293234, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.286950] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293233, 'name': ReconfigVM_Task, 'duration_secs': 0.724283} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.287290] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 0ebb226a-40a3-40cd-8492-ccce87baf0e6/0ebb226a-40a3-40cd-8492-ccce87baf0e6.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 638.287893] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-307b2ebd-99d1-4692-9169-159c604840f7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.298359] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 638.298359] env[62814]: value = "task-4293235" [ 638.298359] env[62814]: _type = "Task" [ 638.298359] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.307269] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293235, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.378124] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dbaee4-8c5b-9fbe-22e3-1c7aa2a24822, 'name': SearchDatastore_Task, 'duration_secs': 0.022415} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.378124] env[62814]: DEBUG nova.compute.manager [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 638.378871] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.379161] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 638.379837] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.379837] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.379837] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 638.379985] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-30b6d8f4-41ed-44f2-9c3c-ede13bc4a144 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.389850] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 638.390072] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 638.390766] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e14293c0-664a-4e15-8300-4eb65753bf6e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.398338] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 638.398338] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526b98aa-47e9-719c-08ed-bc3552b1be44" [ 638.398338] env[62814]: _type = "Task" [ 638.398338] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.411567] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526b98aa-47e9-719c-08ed-bc3552b1be44, 'name': SearchDatastore_Task, 'duration_secs': 0.01056} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.415952] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e0e97cf-928c-407c-b4b6-a190f13b1a1e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.421669] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquiring lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.421902] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.422187] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquiring lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.422354] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 638.422522] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 638.424442] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 638.424442] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520dac0f-7448-bd1b-3fb4-20ebd8b046f2" [ 638.424442] env[62814]: _type = "Task" [ 638.424442] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.431539] env[62814]: INFO nova.compute.manager [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Terminating instance [ 638.441293] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520dac0f-7448-bd1b-3fb4-20ebd8b046f2, 'name': SearchDatastore_Task, 'duration_secs': 0.010308} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.442500] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.442500] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 6a592192-1b41-4be2-84a6-c3b76a4e5643/6a592192-1b41-4be2-84a6-c3b76a4e5643.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 638.442500] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-618f9946-e32c-4226-8b27-3ee56db2dae5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.450086] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 638.450086] env[62814]: value = "task-4293236" [ 638.450086] env[62814]: _type = "Task" [ 638.450086] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.460674] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293236, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.514532] env[62814]: DEBUG nova.compute.manager [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 638.555199] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 32f1bd48-3a2e-42fe-8033-85d3ad55affe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.555356] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 37649169-66e6-4ef1-a4cd-c65b9831004a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.555478] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 591e4a92-7c03-4d70-885f-49a66032c97e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.555597] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 66db614c-d101-409d-a5c3-9de38e5a6eff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.555710] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance be02bc93-1278-4b3d-afa3-270c84585d1d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.555821] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance f5504590-823d-4c37-8226-26684f7957a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.555951] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 0ebb226a-40a3-40cd-8492-ccce87baf0e6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.556082] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 6a592192-1b41-4be2-84a6-c3b76a4e5643 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.556199] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 17bfe703-ff96-4cb7-8535-eed57e10d673 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.556334] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 62bc755d-4f96-4486-884b-0d0c337267aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.556442] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 5087b202-9bba-4489-823b-5d93cbf116e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 638.638358] env[62814]: DEBUG nova.network.neutron [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Updating instance_info_cache with network_info: [{"id": "6e1ae96b-a9fd-4e11-9803-191a35a77cb2", "address": "fa:16:3e:3c:b9:dc", "network": {"id": "b16d5204-f140-44fb-b4c4-621c3f5a3423", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-309109771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5f0190bd5f24a909c9839b7fd1e870b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec46b14d-3310-4f2b-96c1-f53ee47d3759", "external-id": "nsx-vlan-transportzone-323", "segmentation_id": 323, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e1ae96b-a9", "ovs_interfaceid": "6e1ae96b-a9fd-4e11-9803-191a35a77cb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.766696] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293234, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.810480] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293235, 'name': Rename_Task, 'duration_secs': 0.21966} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.810693] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 638.811016] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc2d9657-a376-41e0-bb69-b5131e574085 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.823849] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 638.823849] env[62814]: value = "task-4293237" [ 638.823849] env[62814]: _type = "Task" [ 638.823849] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.833714] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.845294] env[62814]: DEBUG nova.network.neutron [-] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.918072] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 638.938701] env[62814]: DEBUG nova.compute.manager [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 638.939430] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 638.943064] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8cb57fc-0aa2-4ce3-bb82-6920db8951a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.949173] env[62814]: DEBUG nova.compute.manager [req-1481b84e-d3c1-4ea9-8002-9446a3acba55 req-1f48639f-6487-455c-bc8c-b6177c5bd1e2 service nova] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Received event network-changed-cdb5c7dc-7e71-4338-af80-0211db859c91 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 638.949173] env[62814]: DEBUG nova.compute.manager [req-1481b84e-d3c1-4ea9-8002-9446a3acba55 req-1f48639f-6487-455c-bc8c-b6177c5bd1e2 service nova] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Refreshing instance network info cache due to event network-changed-cdb5c7dc-7e71-4338-af80-0211db859c91. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 638.949173] env[62814]: DEBUG oslo_concurrency.lockutils [req-1481b84e-d3c1-4ea9-8002-9446a3acba55 req-1f48639f-6487-455c-bc8c-b6177c5bd1e2 service nova] Acquiring lock "refresh_cache-0ebb226a-40a3-40cd-8492-ccce87baf0e6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.951526] env[62814]: DEBUG oslo_concurrency.lockutils [req-1481b84e-d3c1-4ea9-8002-9446a3acba55 req-1f48639f-6487-455c-bc8c-b6177c5bd1e2 service nova] Acquired lock "refresh_cache-0ebb226a-40a3-40cd-8492-ccce87baf0e6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.951729] env[62814]: DEBUG nova.network.neutron [req-1481b84e-d3c1-4ea9-8002-9446a3acba55 req-1f48639f-6487-455c-bc8c-b6177c5bd1e2 service nova] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Refreshing network info cache for port cdb5c7dc-7e71-4338-af80-0211db859c91 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 638.961329] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 638.961869] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a64e79c9-2025-45e5-976c-afbe42ce84bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.969737] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293236, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.974809] env[62814]: DEBUG oslo_vmware.api [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for the task: (returnval){ [ 638.974809] env[62814]: value = "task-4293238" [ 638.974809] env[62814]: _type = "Task" [ 638.974809] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.988951] env[62814]: DEBUG oslo_vmware.api [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293238, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.062203] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance c7cef7f3-11db-44e1-a454-98830b465b52 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.141133] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Releasing lock "refresh_cache-17bfe703-ff96-4cb7-8535-eed57e10d673" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 639.143757] env[62814]: DEBUG nova.compute.manager [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Instance network_info: |[{"id": "6e1ae96b-a9fd-4e11-9803-191a35a77cb2", "address": "fa:16:3e:3c:b9:dc", "network": {"id": "b16d5204-f140-44fb-b4c4-621c3f5a3423", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-309109771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5f0190bd5f24a909c9839b7fd1e870b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec46b14d-3310-4f2b-96c1-f53ee47d3759", "external-id": "nsx-vlan-transportzone-323", "segmentation_id": 323, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e1ae96b-a9", "ovs_interfaceid": "6e1ae96b-a9fd-4e11-9803-191a35a77cb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 639.143848] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:b9:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ec46b14d-3310-4f2b-96c1-f53ee47d3759', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e1ae96b-a9fd-4e11-9803-191a35a77cb2', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 639.151902] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Creating folder: Project (d5f0190bd5f24a909c9839b7fd1e870b). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.152936] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6de731d8-fd5a-40d1-8e9c-910b06eadff3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.163249] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Created folder: Project (d5f0190bd5f24a909c9839b7fd1e870b) in parent group-v845547. [ 639.163614] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Creating folder: Instances. Parent ref: group-v845572. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 639.164197] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f943f28-aac6-4a4c-8ce1-8ba4f4dd780d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.174695] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Created folder: Instances in parent group-v845572. [ 639.175644] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 639.176195] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 639.176493] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64c4175a-267e-459c-b94c-5264f4689e98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.203283] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 639.203283] env[62814]: value = "task-4293241" [ 639.203283] env[62814]: _type = "Task" [ 639.203283] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.215395] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293241, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.259537] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293234, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.339644] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293237, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.352710] env[62814]: INFO nova.compute.manager [-] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Took 1.85 seconds to deallocate network for instance. [ 639.468590] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293236, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.663222} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.468590] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 6a592192-1b41-4be2-84a6-c3b76a4e5643/6a592192-1b41-4be2-84a6-c3b76a4e5643.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 639.468908] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 639.469084] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a8e136d4-d4d2-4ca1-a201-8968c74e9572 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.476815] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 639.476815] env[62814]: value = "task-4293242" [ 639.476815] env[62814]: _type = "Task" [ 639.476815] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.495624] env[62814]: DEBUG oslo_vmware.api [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293238, 'name': PowerOffVM_Task, 'duration_secs': 0.384785} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.496102] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293242, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.496396] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 639.496597] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 639.496863] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9040f4c4-7d1e-4431-aed7-60de9caa3b0d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.529362] env[62814]: DEBUG nova.compute.manager [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 639.566487] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:34:26Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1847523081',id=20,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1667731819',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 639.566487] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 639.566487] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 639.566738] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 639.566738] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 639.566900] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 639.567194] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 639.567429] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 639.567933] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 639.568291] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 639.568569] env[62814]: DEBUG nova.virt.hardware [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 639.569395] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 3d3f32fc-276a-49be-b471-01a5d6fc5069 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.571472] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d30618d-4385-4ec4-828b-d856f401efe7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.582891] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77c47f89-f6b7-4cb8-b5e1-b415db324063 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.590802] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 639.591364] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 639.591628] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Deleting the datastore file [datastore2] 32f1bd48-3a2e-42fe-8033-85d3ad55affe {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 639.592182] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a64fb46-d611-4e6e-b187-c1694eecc3c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.610385] env[62814]: DEBUG oslo_vmware.api [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for the task: (returnval){ [ 639.610385] env[62814]: value = "task-4293244" [ 639.610385] env[62814]: _type = "Task" [ 639.610385] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.620246] env[62814]: DEBUG oslo_vmware.api [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293244, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.716201] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293241, 'name': CreateVM_Task, 'duration_secs': 0.479155} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.716770] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 639.717170] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.717358] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.717704] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 639.717956] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d218c2d0-f90b-4a22-8bd4-1274046f16fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.726613] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 639.726613] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521b175f-f1ac-876a-b8d3-20ebd69c7855" [ 639.726613] env[62814]: _type = "Task" [ 639.726613] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.740279] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521b175f-f1ac-876a-b8d3-20ebd69c7855, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.755166] env[62814]: DEBUG oslo_vmware.api [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293234, 'name': PowerOnVM_Task, 'duration_secs': 1.726461} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.755499] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 639.755782] env[62814]: INFO nova.compute.manager [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Took 13.51 seconds to spawn the instance on the hypervisor. [ 639.756021] env[62814]: DEBUG nova.compute.manager [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 639.756956] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ed32caf-f41a-4b8e-a7a6-7ba27fe1ec5e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.840677] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293237, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.861448] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 639.934171] env[62814]: DEBUG nova.network.neutron [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Successfully created port: 80990e09-7766-4c2a-937b-52c703caa252 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 639.968412] env[62814]: DEBUG nova.compute.manager [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Received event network-vif-plugged-827a94b4-5864-4060-bfb5-d0e9d2281332 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 639.968412] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Acquiring lock "6a592192-1b41-4be2-84a6-c3b76a4e5643-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 639.968412] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 639.968412] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 639.968412] env[62814]: DEBUG nova.compute.manager [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] No waiting events found dispatching network-vif-plugged-827a94b4-5864-4060-bfb5-d0e9d2281332 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 639.969474] env[62814]: WARNING nova.compute.manager [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Received unexpected event network-vif-plugged-827a94b4-5864-4060-bfb5-d0e9d2281332 for instance with vm_state building and task_state spawning. [ 639.969841] env[62814]: DEBUG nova.compute.manager [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Received event network-changed-827a94b4-5864-4060-bfb5-d0e9d2281332 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 639.970167] env[62814]: DEBUG nova.compute.manager [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Refreshing instance network info cache due to event network-changed-827a94b4-5864-4060-bfb5-d0e9d2281332. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 639.970454] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Acquiring lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.970698] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Acquired lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.970944] env[62814]: DEBUG nova.network.neutron [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Refreshing network info cache for port 827a94b4-5864-4060-bfb5-d0e9d2281332 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 639.989780] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293242, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073568} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.990733] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 639.991614] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2264607b-0afc-4676-b628-e997d3341970 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.021146] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Reconfiguring VM instance instance-00000008 to attach disk [datastore2] 6a592192-1b41-4be2-84a6-c3b76a4e5643/6a592192-1b41-4be2-84a6-c3b76a4e5643.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 640.024256] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78ff6786-f05f-4ee8-a336-ec23b57cfe9f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.046880] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 640.046880] env[62814]: value = "task-4293245" [ 640.046880] env[62814]: _type = "Task" [ 640.046880] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.056669] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293245, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.076625] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 6976b964-a8d3-4886-8aac-8d513e721018 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.123073] env[62814]: DEBUG oslo_vmware.api [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Task: {'id': task-4293244, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.258952} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.123073] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 640.123073] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 640.123073] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 640.123073] env[62814]: INFO nova.compute.manager [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Took 1.18 seconds to destroy the instance on the hypervisor. [ 640.123392] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 640.123392] env[62814]: DEBUG nova.compute.manager [-] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 640.123392] env[62814]: DEBUG nova.network.neutron [-] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 640.239778] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521b175f-f1ac-876a-b8d3-20ebd69c7855, 'name': SearchDatastore_Task, 'duration_secs': 0.014952} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.240113] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.241038] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 640.241038] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.241038] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 640.241038] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 640.241573] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d91a5da5-f9d2-4918-af80-aecf52cc322a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.250899] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 640.251091] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 640.251872] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f75afae-9eb2-4b13-bdef-2e6a06c4b64b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.258429] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 640.258429] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e4a365-2d52-4887-82e3-76f3010c04c2" [ 640.258429] env[62814]: _type = "Task" [ 640.258429] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.267510] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e4a365-2d52-4887-82e3-76f3010c04c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.279807] env[62814]: INFO nova.compute.manager [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Took 26.36 seconds to build instance. [ 640.339684] env[62814]: DEBUG oslo_vmware.api [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293237, 'name': PowerOnVM_Task, 'duration_secs': 1.425067} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.339684] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 640.339684] env[62814]: INFO nova.compute.manager [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Took 10.47 seconds to spawn the instance on the hypervisor. [ 640.339684] env[62814]: DEBUG nova.compute.manager [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 640.339684] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada6b470-7649-433f-ba47-f68eeac836d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.491161] env[62814]: DEBUG nova.network.neutron [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Successfully updated port: 559634c5-106d-4de8-b7b2-596fd81a0c30 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 640.564327] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293245, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.586499] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 296f6c11-7108-42e6-8ada-5d8c08b00da6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.767949] env[62814]: DEBUG nova.network.neutron [req-1481b84e-d3c1-4ea9-8002-9446a3acba55 req-1f48639f-6487-455c-bc8c-b6177c5bd1e2 service nova] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Updated VIF entry in instance network info cache for port cdb5c7dc-7e71-4338-af80-0211db859c91. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 640.772022] env[62814]: DEBUG nova.network.neutron [req-1481b84e-d3c1-4ea9-8002-9446a3acba55 req-1f48639f-6487-455c-bc8c-b6177c5bd1e2 service nova] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Updating instance_info_cache with network_info: [{"id": "cdb5c7dc-7e71-4338-af80-0211db859c91", "address": "fa:16:3e:d3:01:ec", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.123", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdb5c7dc-7e", "ovs_interfaceid": "cdb5c7dc-7e71-4338-af80-0211db859c91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.778965] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e4a365-2d52-4887-82e3-76f3010c04c2, 'name': SearchDatastore_Task, 'duration_secs': 0.034398} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.780253] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47775b9e-903d-416e-9c40-470cc2b0b3a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.785780] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40f98fef-989c-4ca0-b66f-e0af13d13f3d tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lock "f5504590-823d-4c37-8226-26684f7957a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.876s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 640.792250] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 640.792250] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c4a361-db62-d4ea-21b4-92898ca89078" [ 640.792250] env[62814]: _type = "Task" [ 640.792250] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.802343] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c4a361-db62-d4ea-21b4-92898ca89078, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.863712] env[62814]: INFO nova.compute.manager [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Took 26.73 seconds to build instance. [ 640.999483] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "refresh_cache-62bc755d-4f96-4486-884b-0d0c337267aa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.999676] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "refresh_cache-62bc755d-4f96-4486-884b-0d0c337267aa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.001308] env[62814]: DEBUG nova.network.neutron [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 641.061664] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293245, 'name': ReconfigVM_Task, 'duration_secs': 0.784031} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.064056] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Reconfigured VM instance instance-00000008 to attach disk [datastore2] 6a592192-1b41-4be2-84a6-c3b76a4e5643/6a592192-1b41-4be2-84a6-c3b76a4e5643.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 641.064056] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b687308d-5b71-43d2-ad5c-b4102a947a17 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.069634] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 641.069634] env[62814]: value = "task-4293246" [ 641.069634] env[62814]: _type = "Task" [ 641.069634] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.078342] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293246, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.091176] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d9242042-6209-4b04-bf00-00dd04d9d6a0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.091176] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 641.091368] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=149GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 641.237952] env[62814]: DEBUG nova.network.neutron [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updated VIF entry in instance network info cache for port 827a94b4-5864-4060-bfb5-d0e9d2281332. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 641.238469] env[62814]: DEBUG nova.network.neutron [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance_info_cache with network_info: [{"id": "827a94b4-5864-4060-bfb5-d0e9d2281332", "address": "fa:16:3e:a9:8f:95", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.43", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap827a94b4-58", "ovs_interfaceid": "827a94b4-5864-4060-bfb5-d0e9d2281332", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.279164] env[62814]: DEBUG oslo_concurrency.lockutils [req-1481b84e-d3c1-4ea9-8002-9446a3acba55 req-1f48639f-6487-455c-bc8c-b6177c5bd1e2 service nova] Releasing lock "refresh_cache-0ebb226a-40a3-40cd-8492-ccce87baf0e6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.289595] env[62814]: DEBUG nova.compute.manager [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 641.318877] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c4a361-db62-d4ea-21b4-92898ca89078, 'name': SearchDatastore_Task, 'duration_secs': 0.012096} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.318877] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.318877] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 17bfe703-ff96-4cb7-8535-eed57e10d673/17bfe703-ff96-4cb7-8535-eed57e10d673.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 641.319219] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e26883c4-819b-4b55-b04d-3f13554db39b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.325185] env[62814]: DEBUG nova.network.neutron [-] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.331877] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 641.331877] env[62814]: value = "task-4293247" [ 641.331877] env[62814]: _type = "Task" [ 641.331877] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.343208] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293247, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.366127] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f6232243-e303-4e8a-8089-6def3f35c979 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.247s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 641.447026] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0655a978-988b-478a-b85d-f804a5be5e2d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.458544] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc2843f-d644-4fbe-ace2-d16722bf72a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.518314] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4423b493-0e02-4937-8e13-4f6ce6070626 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.529838] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f82975b-4691-44d5-8206-3d0741b4ab6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.545259] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.582898] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293246, 'name': Rename_Task, 'duration_secs': 0.270691} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.583245] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 641.583511] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f99dee1-79a7-4c6c-bf20-d003f2331c13 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.590525] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 641.590525] env[62814]: value = "task-4293248" [ 641.590525] env[62814]: _type = "Task" [ 641.590525] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.599831] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293248, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.749171] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Releasing lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.749171] env[62814]: DEBUG nova.compute.manager [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Received event network-vif-plugged-6e1ae96b-a9fd-4e11-9803-191a35a77cb2 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 641.749381] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Acquiring lock "17bfe703-ff96-4cb7-8535-eed57e10d673-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 641.749519] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Lock "17bfe703-ff96-4cb7-8535-eed57e10d673-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 641.749705] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Lock "17bfe703-ff96-4cb7-8535-eed57e10d673-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 641.753020] env[62814]: DEBUG nova.compute.manager [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] No waiting events found dispatching network-vif-plugged-6e1ae96b-a9fd-4e11-9803-191a35a77cb2 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 641.753020] env[62814]: WARNING nova.compute.manager [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Received unexpected event network-vif-plugged-6e1ae96b-a9fd-4e11-9803-191a35a77cb2 for instance with vm_state building and task_state spawning. [ 641.753020] env[62814]: DEBUG nova.compute.manager [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Received event network-changed-6e1ae96b-a9fd-4e11-9803-191a35a77cb2 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 641.753020] env[62814]: DEBUG nova.compute.manager [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Refreshing instance network info cache due to event network-changed-6e1ae96b-a9fd-4e11-9803-191a35a77cb2. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 641.753020] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Acquiring lock "refresh_cache-17bfe703-ff96-4cb7-8535-eed57e10d673" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 641.753633] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Acquired lock "refresh_cache-17bfe703-ff96-4cb7-8535-eed57e10d673" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.753633] env[62814]: DEBUG nova.network.neutron [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Refreshing network info cache for port 6e1ae96b-a9fd-4e11-9803-191a35a77cb2 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 641.828958] env[62814]: INFO nova.compute.manager [-] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Took 1.71 seconds to deallocate network for instance. [ 641.846835] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 641.853355] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293247, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.886366] env[62814]: DEBUG nova.network.neutron [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.048711] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 642.110089] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293248, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.345459] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293247, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.834603} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.345780] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 17bfe703-ff96-4cb7-8535-eed57e10d673/17bfe703-ff96-4cb7-8535-eed57e10d673.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 642.345914] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 642.346267] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56b91fa2-7fef-4d44-a5ff-134ee487acb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.349851] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.356184] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 642.356184] env[62814]: value = "task-4293249" [ 642.356184] env[62814]: _type = "Task" [ 642.356184] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.366692] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293249, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.389218] env[62814]: DEBUG nova.network.neutron [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Updating instance_info_cache with network_info: [{"id": "559634c5-106d-4de8-b7b2-596fd81a0c30", "address": "fa:16:3e:df:9d:41", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559634c5-10", "ovs_interfaceid": "559634c5-106d-4de8-b7b2-596fd81a0c30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.558343] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 642.558343] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.069s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.558343] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.598s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.558773] env[62814]: INFO nova.compute.claims [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.605331] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293248, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.872420] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "98d134b4-b4ca-4247-a638-ad5c24a694e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.872420] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "98d134b4-b4ca-4247-a638-ad5c24a694e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.879436] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293249, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137874} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.880125] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 642.881472] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec3e782-bda5-4d06-8145-f95c15f97288 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.905546] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "refresh_cache-62bc755d-4f96-4486-884b-0d0c337267aa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.906034] env[62814]: DEBUG nova.compute.manager [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Instance network_info: |[{"id": "559634c5-106d-4de8-b7b2-596fd81a0c30", "address": "fa:16:3e:df:9d:41", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559634c5-10", "ovs_interfaceid": "559634c5-106d-4de8-b7b2-596fd81a0c30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 642.922962] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Reconfiguring VM instance instance-00000009 to attach disk [datastore2] 17bfe703-ff96-4cb7-8535-eed57e10d673/17bfe703-ff96-4cb7-8535-eed57e10d673.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 642.927957] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:9d:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3db2ab9e-1244-4377-b05f-ab76003f2428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '559634c5-106d-4de8-b7b2-596fd81a0c30', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 642.940762] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating folder: Project (3dedbeff832a4ac48b0aa01bd6acc3f1). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 642.941772] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fde98a72-2454-4a56-8671-48bf17d2334a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.967805] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9089fc75-a1c8-42df-9e72-8fb5cc0685a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.977146] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 642.977146] env[62814]: value = "task-4293250" [ 642.977146] env[62814]: _type = "Task" [ 642.977146] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.984392] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Created folder: Project (3dedbeff832a4ac48b0aa01bd6acc3f1) in parent group-v845547. [ 642.984662] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating folder: Instances. Parent ref: group-v845575. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 642.985503] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a2e35005-5f54-4a1a-9f08-222978607616 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.993497] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293250, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.002994] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Created folder: Instances in parent group-v845575. [ 643.003385] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 643.004138] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 643.004138] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97369fe6-a3ab-4721-a232-e80e9a007e29 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.024334] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 643.024334] env[62814]: value = "task-4293253" [ 643.024334] env[62814]: _type = "Task" [ 643.024334] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.032777] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293253, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.064141] env[62814]: DEBUG nova.network.neutron [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Updated VIF entry in instance network info cache for port 6e1ae96b-a9fd-4e11-9803-191a35a77cb2. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 643.064479] env[62814]: DEBUG nova.network.neutron [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Updating instance_info_cache with network_info: [{"id": "6e1ae96b-a9fd-4e11-9803-191a35a77cb2", "address": "fa:16:3e:3c:b9:dc", "network": {"id": "b16d5204-f140-44fb-b4c4-621c3f5a3423", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-309109771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5f0190bd5f24a909c9839b7fd1e870b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec46b14d-3310-4f2b-96c1-f53ee47d3759", "external-id": "nsx-vlan-transportzone-323", "segmentation_id": 323, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e1ae96b-a9", "ovs_interfaceid": "6e1ae96b-a9fd-4e11-9803-191a35a77cb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.106917] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293248, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.343127] env[62814]: DEBUG nova.network.neutron [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Successfully updated port: 80990e09-7766-4c2a-937b-52c703caa252 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 643.381285] env[62814]: DEBUG nova.compute.manager [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 643.408136] env[62814]: DEBUG nova.compute.manager [req-5af2c7c5-3b20-47a2-a9fb-f7c8e8a58de2 req-9dd41300-a172-402f-8b65-0f843dcf5119 service nova] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Received event network-vif-deleted-5dc7f23e-f7a5-49d3-8451-88a7f80d7c90 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 643.408440] env[62814]: DEBUG nova.compute.manager [req-5af2c7c5-3b20-47a2-a9fb-f7c8e8a58de2 req-9dd41300-a172-402f-8b65-0f843dcf5119 service nova] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Received event network-vif-deleted-a896bdd8-35a5-4767-a971-88307c117164 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 643.490305] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293250, 'name': ReconfigVM_Task, 'duration_secs': 0.38322} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.490579] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Reconfigured VM instance instance-00000009 to attach disk [datastore2] 17bfe703-ff96-4cb7-8535-eed57e10d673/17bfe703-ff96-4cb7-8535-eed57e10d673.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 643.491213] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e25a8a0-dd90-4652-b779-edf904649dda {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.501928] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 643.501928] env[62814]: value = "task-4293254" [ 643.501928] env[62814]: _type = "Task" [ 643.501928] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.514294] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293254, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.537229] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293253, 'name': CreateVM_Task, 'duration_secs': 0.386357} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.537229] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 643.537229] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.537229] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 643.537229] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 643.537478] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcf8f2eb-3b01-443d-8258-0b3c0b761dc8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.543898] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 643.543898] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527f7440-263e-e54d-baf4-1558536602de" [ 643.543898] env[62814]: _type = "Task" [ 643.543898] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.552659] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527f7440-263e-e54d-baf4-1558536602de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.570583] env[62814]: DEBUG oslo_concurrency.lockutils [req-e793ae4c-1648-443a-95a6-88d59f5b5401 req-d25e286d-7801-4285-bce0-74dc97399778 service nova] Releasing lock "refresh_cache-17bfe703-ff96-4cb7-8535-eed57e10d673" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.606470] env[62814]: DEBUG oslo_vmware.api [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293248, 'name': PowerOnVM_Task, 'duration_secs': 1.924166} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.607086] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 643.607086] env[62814]: INFO nova.compute.manager [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Took 11.36 seconds to spawn the instance on the hypervisor. [ 643.607286] env[62814]: DEBUG nova.compute.manager [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 643.608015] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1a37ecd-d83a-45de-9624-535c585c1798 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.845236] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "refresh_cache-5087b202-9bba-4489-823b-5d93cbf116e2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.845644] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquired lock "refresh_cache-5087b202-9bba-4489-823b-5d93cbf116e2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 643.845774] env[62814]: DEBUG nova.network.neutron [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.907355] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.939563] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15fccd31-c80b-4640-bb87-c4ec4e4ecf65 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.950412] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e89d41-fd19-4861-a937-21e1b379aede {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.986275] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6182ca67-a955-4be7-94c1-f762c4128bb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.993556] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819b125d-07b7-4196-9196-acbd8f4b81ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.007734] env[62814]: DEBUG nova.compute.provider_tree [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 644.017801] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293254, 'name': Rename_Task, 'duration_secs': 0.225033} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.018857] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 644.019128] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed0fccae-9f25-42df-9492-aa91f4c8416b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.025703] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 644.025703] env[62814]: value = "task-4293255" [ 644.025703] env[62814]: _type = "Task" [ 644.025703] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.036735] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293255, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.053056] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527f7440-263e-e54d-baf4-1558536602de, 'name': SearchDatastore_Task, 'duration_secs': 0.017898} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.053673] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 644.053673] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 644.053895] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.054130] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 644.054414] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 644.054904] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-87b75310-26f0-41d9-9448-68f2140b7f01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.064728] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 644.064915] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 644.065664] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7af75aa-214e-478a-bfc2-ee9525e65f0d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.072151] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 644.072151] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d5d5bc-c1b0-813e-23c0-655f6da4a753" [ 644.072151] env[62814]: _type = "Task" [ 644.072151] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.080865] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d5d5bc-c1b0-813e-23c0-655f6da4a753, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.139443] env[62814]: INFO nova.compute.manager [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Took 29.23 seconds to build instance. [ 644.366917] env[62814]: DEBUG nova.compute.manager [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Received event network-vif-plugged-559634c5-106d-4de8-b7b2-596fd81a0c30 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 644.367205] env[62814]: DEBUG oslo_concurrency.lockutils [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] Acquiring lock "62bc755d-4f96-4486-884b-0d0c337267aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.367451] env[62814]: DEBUG oslo_concurrency.lockutils [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] Lock "62bc755d-4f96-4486-884b-0d0c337267aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.367672] env[62814]: DEBUG oslo_concurrency.lockutils [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] Lock "62bc755d-4f96-4486-884b-0d0c337267aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 644.368106] env[62814]: DEBUG nova.compute.manager [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] No waiting events found dispatching network-vif-plugged-559634c5-106d-4de8-b7b2-596fd81a0c30 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 644.368363] env[62814]: WARNING nova.compute.manager [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Received unexpected event network-vif-plugged-559634c5-106d-4de8-b7b2-596fd81a0c30 for instance with vm_state building and task_state spawning. [ 644.368543] env[62814]: DEBUG nova.compute.manager [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Received event network-changed-559634c5-106d-4de8-b7b2-596fd81a0c30 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 644.368696] env[62814]: DEBUG nova.compute.manager [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Refreshing instance network info cache due to event network-changed-559634c5-106d-4de8-b7b2-596fd81a0c30. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 644.368890] env[62814]: DEBUG oslo_concurrency.lockutils [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] Acquiring lock "refresh_cache-62bc755d-4f96-4486-884b-0d0c337267aa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.369039] env[62814]: DEBUG oslo_concurrency.lockutils [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] Acquired lock "refresh_cache-62bc755d-4f96-4486-884b-0d0c337267aa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 644.369237] env[62814]: DEBUG nova.network.neutron [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Refreshing network info cache for port 559634c5-106d-4de8-b7b2-596fd81a0c30 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 644.416778] env[62814]: DEBUG nova.network.neutron [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.538960] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293255, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.544191] env[62814]: ERROR nova.scheduler.client.report [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [req-b8f753c7-d03e-4ef6-80c7-f8b82d144870] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b8f753c7-d03e-4ef6-80c7-f8b82d144870"}]} [ 644.563994] env[62814]: DEBUG nova.scheduler.client.report [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 644.579982] env[62814]: DEBUG nova.scheduler.client.report [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 644.580438] env[62814]: DEBUG nova.compute.provider_tree [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 644.590992] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d5d5bc-c1b0-813e-23c0-655f6da4a753, 'name': SearchDatastore_Task, 'duration_secs': 0.017821} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.591912] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f3c422d-5256-413a-801b-4b90d18d3de3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.597123] env[62814]: DEBUG nova.scheduler.client.report [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 644.601116] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 644.601116] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522dfc8c-09a5-ce63-6ff2-672f53e02ee8" [ 644.601116] env[62814]: _type = "Task" [ 644.601116] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.613329] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522dfc8c-09a5-ce63-6ff2-672f53e02ee8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.627026] env[62814]: DEBUG nova.scheduler.client.report [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 644.642530] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18a4253f-96fa-4862-9bf5-6f3de789cea9 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.746s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.027035] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c100f66f-9029-46d7-901f-e819b484aaa9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.054824] env[62814]: DEBUG oslo_vmware.api [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293255, 'name': PowerOnVM_Task, 'duration_secs': 0.87617} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.063808] env[62814]: DEBUG nova.network.neutron [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Updating instance_info_cache with network_info: [{"id": "80990e09-7766-4c2a-937b-52c703caa252", "address": "fa:16:3e:92:a4:40", "network": {"id": "b4436b99-a4a3-46c9-918a-4e6f6f6bb55e", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-2020231604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d51711cd907404b8d0a48d904b599c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98d96b75-ac36-499a-adc2-130c8c1d55ca", "external-id": "nsx-vlan-transportzone-564", "segmentation_id": 564, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80990e09-77", "ovs_interfaceid": "80990e09-7766-4c2a-937b-52c703caa252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.066085] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e2e725-b33b-4238-9b00-fbc82c6e07e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.075651] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 645.075865] env[62814]: INFO nova.compute.manager [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Took 10.40 seconds to spawn the instance on the hypervisor. [ 645.076050] env[62814]: DEBUG nova.compute.manager [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 645.078307] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Acquiring lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.078515] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.078703] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Acquiring lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.078871] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.079067] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.081302] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1794f666-6263-45d5-bb83-5e6d0385f5dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.085109] env[62814]: DEBUG nova.compute.manager [None req-3ab5e90c-c91e-443d-9eb5-deff962525cd tempest-ServerDiagnosticsV248Test-1633039967 tempest-ServerDiagnosticsV248Test-1633039967-project-admin] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 645.086355] env[62814]: INFO nova.compute.manager [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Terminating instance [ 645.088683] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279298eb-eea0-448c-b252-bbdd89d1ad00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.126194] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6e22ea-3134-489f-8234-0513d3efaafa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.135420] env[62814]: INFO nova.compute.manager [None req-3ab5e90c-c91e-443d-9eb5-deff962525cd tempest-ServerDiagnosticsV248Test-1633039967 tempest-ServerDiagnosticsV248Test-1633039967-project-admin] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Retrieving diagnostics [ 645.140508] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7877f45-1a79-4196-a3fa-4d3bbcc49d5b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.149467] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522dfc8c-09a5-ce63-6ff2-672f53e02ee8, 'name': SearchDatastore_Task, 'duration_secs': 0.019793} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.150063] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 645.150318] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 62bc755d-4f96-4486-884b-0d0c337267aa/62bc755d-4f96-4486-884b-0d0c337267aa.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 645.150904] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d73ec4b-dad5-49f6-bfeb-da968e04df33 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.182112] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-568ea3c1-7a05-4cb6-8c2a-f354f0521d3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.187891] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 645.187891] env[62814]: value = "task-4293256" [ 645.187891] env[62814]: _type = "Task" [ 645.187891] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.205036] env[62814]: DEBUG nova.compute.provider_tree [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 645.215669] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293256, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.217556] env[62814]: DEBUG nova.compute.manager [None req-56031af6-ef62-4a4d-aeca-21d10600c95a tempest-ServerDiagnosticsTest-2064212551 tempest-ServerDiagnosticsTest-2064212551-project-admin] [instance: f5504590-823d-4c37-8226-26684f7957a6] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 645.222098] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f68919a7-23af-42a8-b174-a97ccdcf3c77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.231134] env[62814]: INFO nova.compute.manager [None req-56031af6-ef62-4a4d-aeca-21d10600c95a tempest-ServerDiagnosticsTest-2064212551 tempest-ServerDiagnosticsTest-2064212551-project-admin] [instance: f5504590-823d-4c37-8226-26684f7957a6] Retrieving diagnostics [ 645.233634] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe37d5d-aa50-4db8-bc52-7467107adf2c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.415574] env[62814]: DEBUG nova.network.neutron [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Updated VIF entry in instance network info cache for port 559634c5-106d-4de8-b7b2-596fd81a0c30. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 645.415918] env[62814]: DEBUG nova.network.neutron [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Updating instance_info_cache with network_info: [{"id": "559634c5-106d-4de8-b7b2-596fd81a0c30", "address": "fa:16:3e:df:9d:41", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559634c5-10", "ovs_interfaceid": "559634c5-106d-4de8-b7b2-596fd81a0c30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.579217] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Releasing lock "refresh_cache-5087b202-9bba-4489-823b-5d93cbf116e2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 645.579686] env[62814]: DEBUG nova.compute.manager [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Instance network_info: |[{"id": "80990e09-7766-4c2a-937b-52c703caa252", "address": "fa:16:3e:92:a4:40", "network": {"id": "b4436b99-a4a3-46c9-918a-4e6f6f6bb55e", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-2020231604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d51711cd907404b8d0a48d904b599c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98d96b75-ac36-499a-adc2-130c8c1d55ca", "external-id": "nsx-vlan-transportzone-564", "segmentation_id": 564, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80990e09-77", "ovs_interfaceid": "80990e09-7766-4c2a-937b-52c703caa252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 645.580237] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:a4:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98d96b75-ac36-499a-adc2-130c8c1d55ca', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '80990e09-7766-4c2a-937b-52c703caa252', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 645.591761] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Creating folder: Project (9d51711cd907404b8d0a48d904b599c4). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 645.592187] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-00c05d95-11ab-4c34-865d-d8eea5bc2dcb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.603991] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Created folder: Project (9d51711cd907404b8d0a48d904b599c4) in parent group-v845547. [ 645.604221] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Creating folder: Instances. Parent ref: group-v845578. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 645.604464] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0bb85383-a64e-418a-b3d2-03e01c4d7151 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.614922] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Created folder: Instances in parent group-v845578. [ 645.614922] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 645.614922] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 645.614922] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f45fabb4-c65b-4a7a-a206-4419cc7831e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.630810] env[62814]: DEBUG nova.compute.manager [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 645.631453] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 645.632623] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70bbab35-1672-4184-9d95-a9af82bec6ad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.642068] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 645.642407] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96c661b9-db5f-46a2-aa6a-84db391f1c43 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.644349] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 645.644349] env[62814]: value = "task-4293259" [ 645.644349] env[62814]: _type = "Task" [ 645.644349] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.659771] env[62814]: INFO nova.compute.manager [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Took 22.61 seconds to build instance. [ 645.661508] env[62814]: DEBUG oslo_vmware.api [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Waiting for the task: (returnval){ [ 645.661508] env[62814]: value = "task-4293260" [ 645.661508] env[62814]: _type = "Task" [ 645.661508] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.670072] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293259, 'name': CreateVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.681084] env[62814]: DEBUG oslo_vmware.api [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Task: {'id': task-4293260, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.701533] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293256, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.778035] env[62814]: DEBUG nova.scheduler.client.report [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 22 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 645.783316] env[62814]: DEBUG nova.compute.provider_tree [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 22 to 23 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 645.783316] env[62814]: DEBUG nova.compute.provider_tree [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 645.918530] env[62814]: DEBUG oslo_concurrency.lockutils [req-a0109663-7639-4dd6-b9c2-2afd2305c9a7 req-8d5b5edf-2013-4a6a-9fde-cc3761613441 service nova] Releasing lock "refresh_cache-62bc755d-4f96-4486-884b-0d0c337267aa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 646.160391] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293259, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.164515] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9df3d3ab-62fb-488f-a0e8-0358b1805c49 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "17bfe703-ff96-4cb7-8535-eed57e10d673" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.133s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.179125] env[62814]: DEBUG oslo_vmware.api [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Task: {'id': task-4293260, 'name': PowerOffVM_Task, 'duration_secs': 0.36746} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.179627] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 646.179848] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 646.180138] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9df0bb70-fb1c-4418-b34b-3ec23d5f7363 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.182169] env[62814]: DEBUG oslo_concurrency.lockutils [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquiring lock "66db614c-d101-409d-a5c3-9de38e5a6eff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.182362] env[62814]: DEBUG oslo_concurrency.lockutils [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lock "66db614c-d101-409d-a5c3-9de38e5a6eff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.182774] env[62814]: DEBUG oslo_concurrency.lockutils [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquiring lock "66db614c-d101-409d-a5c3-9de38e5a6eff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.182984] env[62814]: DEBUG oslo_concurrency.lockutils [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lock "66db614c-d101-409d-a5c3-9de38e5a6eff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.183166] env[62814]: DEBUG oslo_concurrency.lockutils [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lock "66db614c-d101-409d-a5c3-9de38e5a6eff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.186508] env[62814]: INFO nova.compute.manager [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Terminating instance [ 646.202578] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293256, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.795204} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.203546] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 62bc755d-4f96-4486-884b-0d0c337267aa/62bc755d-4f96-4486-884b-0d0c337267aa.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 646.203902] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 646.204785] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6297b69e-e123-4370-8f92-17268b23638e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.212134] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 646.212134] env[62814]: value = "task-4293266" [ 646.212134] env[62814]: _type = "Task" [ 646.212134] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.222982] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293266, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.276243] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 646.279145] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 646.279145] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Deleting the datastore file [datastore2] 0ebb226a-40a3-40cd-8492-ccce87baf0e6 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 646.279145] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd53c9a7-cd8a-4722-9475-935aead2bb92 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.287825] env[62814]: DEBUG oslo_vmware.api [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Waiting for the task: (returnval){ [ 646.287825] env[62814]: value = "task-4293267" [ 646.287825] env[62814]: _type = "Task" [ 646.287825] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.293590] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.736s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.296035] env[62814]: DEBUG nova.compute.manager [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 646.298717] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.074s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.302941] env[62814]: INFO nova.compute.claims [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.313904] env[62814]: DEBUG oslo_vmware.api [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Task: {'id': task-4293267, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.654354] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293259, 'name': CreateVM_Task, 'duration_secs': 0.645799} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.654599] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 646.658122] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.658122] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 646.658122] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 646.658122] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff2f878f-bbac-4067-8e16-225d58209ee5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.662702] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 646.662702] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524066fa-e6eb-c4ed-e2c0-1f31bbcb9035" [ 646.662702] env[62814]: _type = "Task" [ 646.662702] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.673262] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524066fa-e6eb-c4ed-e2c0-1f31bbcb9035, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.692554] env[62814]: DEBUG nova.compute.manager [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 646.692861] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 646.694509] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6061806-0b1c-4b98-b1c5-c7db5bc752ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.703481] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 646.703635] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c7ec50bc-bcbd-403e-85fd-86217f76c5db {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.709944] env[62814]: DEBUG oslo_vmware.api [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for the task: (returnval){ [ 646.709944] env[62814]: value = "task-4293268" [ 646.709944] env[62814]: _type = "Task" [ 646.709944] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.736239] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293266, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075756} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.736575] env[62814]: DEBUG oslo_vmware.api [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293268, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.736847] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 646.738969] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdcdc6d0-6cd6-4daf-a911-c53a71051284 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.762377] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] 62bc755d-4f96-4486-884b-0d0c337267aa/62bc755d-4f96-4486-884b-0d0c337267aa.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 646.762377] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b6eb47a-a7ea-46fb-bc68-409d0bc565b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.785680] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 646.785680] env[62814]: value = "task-4293269" [ 646.785680] env[62814]: _type = "Task" [ 646.785680] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.800735] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.805526] env[62814]: DEBUG oslo_vmware.api [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Task: {'id': task-4293267, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.415676} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.806020] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 646.806020] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 646.806261] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 646.806454] env[62814]: INFO nova.compute.manager [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Took 1.18 seconds to destroy the instance on the hypervisor. [ 646.806691] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 646.806875] env[62814]: DEBUG nova.compute.manager [-] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 646.807436] env[62814]: DEBUG nova.network.neutron [-] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.815029] env[62814]: DEBUG nova.compute.utils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 646.815029] env[62814]: DEBUG nova.compute.manager [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 646.815029] env[62814]: DEBUG nova.network.neutron [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 647.136204] env[62814]: DEBUG nova.policy [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '307242d782d24d1594ee89c0e46e7aa3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42a09e557c794726aa4895c1563ea62f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 647.176937] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524066fa-e6eb-c4ed-e2c0-1f31bbcb9035, 'name': SearchDatastore_Task, 'duration_secs': 0.013238} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.177708] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 647.177708] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 647.177819] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.177987] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 647.178227] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 647.179963] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d398e6e5-b5f3-4b39-b1f7-4bd3b1cc9194 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.191417] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 647.191609] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 647.192466] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a12a9fc-8a7c-44d8-b8dd-d6384be1db22 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.200250] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 647.200250] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c7e22e-ffba-c261-9d3c-b78df11bccd4" [ 647.200250] env[62814]: _type = "Task" [ 647.200250] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.209637] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c7e22e-ffba-c261-9d3c-b78df11bccd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.223443] env[62814]: DEBUG oslo_vmware.api [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293268, 'name': PowerOffVM_Task, 'duration_secs': 0.203814} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.223724] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 647.223904] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 647.224229] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b8dd9b2b-a8de-4c66-85a1-f2d046564d87 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.237241] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.237241] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 647.299463] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293269, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.302277] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 647.302479] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 647.302657] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Deleting the datastore file [datastore2] 66db614c-d101-409d-a5c3-9de38e5a6eff {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 647.304768] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-320e59b8-c162-4163-a7a9-c010dc9fa116 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.311211] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.311211] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 647.318384] env[62814]: DEBUG oslo_vmware.api [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for the task: (returnval){ [ 647.318384] env[62814]: value = "task-4293271" [ 647.318384] env[62814]: _type = "Task" [ 647.318384] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.319556] env[62814]: DEBUG nova.compute.manager [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 647.348181] env[62814]: DEBUG oslo_vmware.api [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293271, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.679844] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d10e6115-3b51-41bf-b267-28b9304c7671 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.691564] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58f1193-cb4a-49c5-85d4-a552b0df1e2c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.739177] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa37e8b-2d34-4f4f-9009-4ea01a88ce7a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.745892] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 647.750399] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c7e22e-ffba-c261-9d3c-b78df11bccd4, 'name': SearchDatastore_Task, 'duration_secs': 0.015761} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.751335] env[62814]: DEBUG nova.network.neutron [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Successfully created port: 22bf5740-3f2e-45ed-be98-ac62c70cfb3b {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 647.756022] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31f30dd5-a019-4042-afad-119c0477ecb2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.761252] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ab3b80-6950-4cd1-87b4-013572b76d09 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.767206] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 647.767206] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524fb895-b111-1984-9723-b8511444a992" [ 647.767206] env[62814]: _type = "Task" [ 647.767206] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.781801] env[62814]: DEBUG nova.compute.provider_tree [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 647.790394] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524fb895-b111-1984-9723-b8511444a992, 'name': SearchDatastore_Task, 'duration_secs': 0.012294} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.795337] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 647.795768] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 5087b202-9bba-4489-823b-5d93cbf116e2/5087b202-9bba-4489-823b-5d93cbf116e2.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 647.796180] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b45b4d6e-3acd-495c-a45b-30d5479fdfcd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.808262] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293269, 'name': ReconfigVM_Task, 'duration_secs': 0.661042} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.810765] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Reconfigured VM instance instance-0000000a to attach disk [datastore2] 62bc755d-4f96-4486-884b-0d0c337267aa/62bc755d-4f96-4486-884b-0d0c337267aa.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 647.811942] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 647.811942] env[62814]: value = "task-4293272" [ 647.811942] env[62814]: _type = "Task" [ 647.811942] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.812770] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e4b97ac-9edf-45a3-b051-20bba44d3d6e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.816196] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 647.834923] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 647.834923] env[62814]: value = "task-4293273" [ 647.834923] env[62814]: _type = "Task" [ 647.834923] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 647.846980] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293272, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 647.859740] env[62814]: DEBUG oslo_vmware.api [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Task: {'id': task-4293271, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323344} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.859740] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 647.859740] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 647.859740] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 647.859740] env[62814]: INFO nova.compute.manager [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Took 1.17 seconds to destroy the instance on the hypervisor. [ 647.860076] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 647.860076] env[62814]: DEBUG nova.compute.manager [-] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 647.860076] env[62814]: DEBUG nova.network.neutron [-] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 647.868226] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293273, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.169332] env[62814]: DEBUG nova.network.neutron [-] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.282226] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.320871] env[62814]: ERROR nova.scheduler.client.report [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [req-08d72bfa-6ea4-4601-9d76-182a342c6517] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-08d72bfa-6ea4-4601-9d76-182a342c6517"}]} [ 648.339911] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293272, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.355765] env[62814]: DEBUG nova.compute.manager [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 648.358213] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293273, 'name': Rename_Task, 'duration_secs': 0.172008} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.359464] env[62814]: DEBUG nova.scheduler.client.report [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 648.365141] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 648.365141] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 648.366077] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b1446ee-9c21-4c44-b4fe-1cebd40e001a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.378508] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 648.378508] env[62814]: value = "task-4293274" [ 648.378508] env[62814]: _type = "Task" [ 648.378508] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.389057] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293274, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.395668] env[62814]: DEBUG nova.scheduler.client.report [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 648.395906] env[62814]: DEBUG nova.compute.provider_tree [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 648.407836] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 648.407836] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 648.408196] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 648.409231] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 648.409498] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 648.409664] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 648.409894] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 648.410120] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 648.410336] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 648.410511] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 648.410687] env[62814]: DEBUG nova.virt.hardware [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 648.413844] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657533d7-bd46-4da1-a7d4-1d5ed5549780 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.417952] env[62814]: DEBUG nova.scheduler.client.report [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 648.426172] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e619e3-bbf9-4bda-beaf-5b1e76ef88b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.450569] env[62814]: DEBUG nova.scheduler.client.report [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 648.547054] env[62814]: DEBUG nova.compute.manager [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Stashing vm_state: active {{(pid=62814) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 648.673279] env[62814]: INFO nova.compute.manager [-] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Took 1.87 seconds to deallocate network for instance. [ 648.827552] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293272, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.666627} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 648.828085] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 5087b202-9bba-4489-823b-5d93cbf116e2/5087b202-9bba-4489-823b-5d93cbf116e2.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 648.828085] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 648.828464] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c21597d6-e2f1-4207-a0f1-aa20f7b7db67 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.840665] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 648.840665] env[62814]: value = "task-4293275" [ 648.840665] env[62814]: _type = "Task" [ 648.840665] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 648.850272] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293275, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.890353] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293274, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 648.912983] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89c797d0-231e-40c5-8a86-21b71ebe0ac2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.924157] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda47cf9-88bc-4754-993b-0bf205e2ca4b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.969365] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2fda63e-6964-4615-9d88-27869d57ec36 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.978144] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6bcd986-5e55-451e-8ccd-440340b8ff29 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.003821] env[62814]: DEBUG nova.compute.provider_tree [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 649.073458] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.154613] env[62814]: DEBUG nova.network.neutron [-] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.186443] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.356615] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293275, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080508} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.356898] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 649.357805] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c825dab8-9594-43ab-8390-53256a6cc66b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.386760] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Reconfiguring VM instance instance-0000000b to attach disk [datastore2] 5087b202-9bba-4489-823b-5d93cbf116e2/5087b202-9bba-4489-823b-5d93cbf116e2.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 649.391051] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-909c6c97-ef98-45f4-9b6a-b699834938ef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.413801] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293274, 'name': PowerOnVM_Task} progress is 91%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.416091] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 649.416091] env[62814]: value = "task-4293277" [ 649.416091] env[62814]: _type = "Task" [ 649.416091] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 649.424426] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293277, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 649.517290] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquiring lock "591e4a92-7c03-4d70-885f-49a66032c97e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.517567] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lock "591e4a92-7c03-4d70-885f-49a66032c97e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 649.517804] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquiring lock "591e4a92-7c03-4d70-885f-49a66032c97e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 649.518703] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lock "591e4a92-7c03-4d70-885f-49a66032c97e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 649.519346] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lock "591e4a92-7c03-4d70-885f-49a66032c97e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 649.522557] env[62814]: INFO nova.compute.manager [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Terminating instance [ 649.563797] env[62814]: DEBUG nova.scheduler.client.report [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 27 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 649.564215] env[62814]: DEBUG nova.compute.provider_tree [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 27 to 28 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 649.564446] env[62814]: DEBUG nova.compute.provider_tree [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 649.660626] env[62814]: INFO nova.compute.manager [-] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Took 1.80 seconds to deallocate network for instance. [ 649.888381] env[62814]: DEBUG oslo_vmware.api [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293274, 'name': PowerOnVM_Task, 'duration_secs': 1.04392} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 649.888733] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 649.888947] env[62814]: INFO nova.compute.manager [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Took 12.81 seconds to spawn the instance on the hypervisor. [ 649.889782] env[62814]: DEBUG nova.compute.manager [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 649.890532] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77187e8-1d0f-4602-ad5c-2b6c4672917d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.932090] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293277, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.027345] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquiring lock "refresh_cache-591e4a92-7c03-4d70-885f-49a66032c97e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.027535] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquired lock "refresh_cache-591e4a92-7c03-4d70-885f-49a66032c97e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.028692] env[62814]: DEBUG nova.network.neutron [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 650.043043] env[62814]: DEBUG oslo_concurrency.lockutils [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquiring lock "f5504590-823d-4c37-8226-26684f7957a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.043043] env[62814]: DEBUG oslo_concurrency.lockutils [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lock "f5504590-823d-4c37-8226-26684f7957a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.043043] env[62814]: DEBUG oslo_concurrency.lockutils [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquiring lock "f5504590-823d-4c37-8226-26684f7957a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.043043] env[62814]: DEBUG oslo_concurrency.lockutils [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lock "f5504590-823d-4c37-8226-26684f7957a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.043390] env[62814]: DEBUG oslo_concurrency.lockutils [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lock "f5504590-823d-4c37-8226-26684f7957a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.047352] env[62814]: INFO nova.compute.manager [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Terminating instance [ 650.070835] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.772s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.071445] env[62814]: DEBUG nova.compute.manager [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 650.075727] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.535s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.079144] env[62814]: INFO nova.compute.claims [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 650.170923] env[62814]: DEBUG oslo_concurrency.lockutils [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.324954] env[62814]: DEBUG nova.network.neutron [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Successfully updated port: 22bf5740-3f2e-45ed-be98-ac62c70cfb3b {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 650.375830] env[62814]: DEBUG nova.compute.manager [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Received event network-vif-plugged-80990e09-7766-4c2a-937b-52c703caa252 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 650.376170] env[62814]: DEBUG oslo_concurrency.lockutils [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] Acquiring lock "5087b202-9bba-4489-823b-5d93cbf116e2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.376956] env[62814]: DEBUG oslo_concurrency.lockutils [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] Lock "5087b202-9bba-4489-823b-5d93cbf116e2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.377216] env[62814]: DEBUG oslo_concurrency.lockutils [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] Lock "5087b202-9bba-4489-823b-5d93cbf116e2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.378457] env[62814]: DEBUG nova.compute.manager [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] No waiting events found dispatching network-vif-plugged-80990e09-7766-4c2a-937b-52c703caa252 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 650.378457] env[62814]: WARNING nova.compute.manager [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Received unexpected event network-vif-plugged-80990e09-7766-4c2a-937b-52c703caa252 for instance with vm_state building and task_state spawning. [ 650.378457] env[62814]: DEBUG nova.compute.manager [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Received event network-changed-80990e09-7766-4c2a-937b-52c703caa252 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 650.378457] env[62814]: DEBUG nova.compute.manager [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Refreshing instance network info cache due to event network-changed-80990e09-7766-4c2a-937b-52c703caa252. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 650.378457] env[62814]: DEBUG oslo_concurrency.lockutils [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] Acquiring lock "refresh_cache-5087b202-9bba-4489-823b-5d93cbf116e2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.378692] env[62814]: DEBUG oslo_concurrency.lockutils [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] Acquired lock "refresh_cache-5087b202-9bba-4489-823b-5d93cbf116e2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.378728] env[62814]: DEBUG nova.network.neutron [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Refreshing network info cache for port 80990e09-7766-4c2a-937b-52c703caa252 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 650.410852] env[62814]: INFO nova.compute.manager [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Took 26.22 seconds to build instance. [ 650.427110] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293277, 'name': ReconfigVM_Task, 'duration_secs': 0.900655} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.427404] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Reconfigured VM instance instance-0000000b to attach disk [datastore2] 5087b202-9bba-4489-823b-5d93cbf116e2/5087b202-9bba-4489-823b-5d93cbf116e2.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 650.428286] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d8bcba96-2d14-440c-b957-9a8e1d2c59da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.435343] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 650.435343] env[62814]: value = "task-4293278" [ 650.435343] env[62814]: _type = "Task" [ 650.435343] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.445770] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293278, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.556742] env[62814]: DEBUG nova.compute.manager [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 650.557012] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 650.558040] env[62814]: DEBUG nova.network.neutron [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.561641] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b849d7-f928-4c13-ac69-c25d22577099 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.570951] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 650.571231] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-30271527-cef6-4891-93ea-692d8a470dcc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.580822] env[62814]: DEBUG nova.compute.utils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 650.586352] env[62814]: DEBUG nova.compute.manager [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 650.586587] env[62814]: DEBUG nova.network.neutron [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 650.588580] env[62814]: DEBUG oslo_vmware.api [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for the task: (returnval){ [ 650.588580] env[62814]: value = "task-4293279" [ 650.588580] env[62814]: _type = "Task" [ 650.588580] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.598696] env[62814]: DEBUG oslo_vmware.api [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293279, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 650.712795] env[62814]: DEBUG nova.policy [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '784b28e585654420a44cd8ef1a21dede', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '07ee864ce77f44de84624ee579fb5316', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 650.716638] env[62814]: DEBUG nova.network.neutron [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.828614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.828614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 650.828754] env[62814]: DEBUG nova.network.neutron [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 650.913251] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0cadc363-b5da-45ee-982e-5715f1ba78ba tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "62bc755d-4f96-4486-884b-0d0c337267aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.744s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.953436] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293278, 'name': Rename_Task, 'duration_secs': 0.277313} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 650.953757] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 650.954048] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02bccf38-7e92-4e82-8b59-ff060171ce5a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.963141] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 650.963141] env[62814]: value = "task-4293280" [ 650.963141] env[62814]: _type = "Task" [ 650.963141] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.975385] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293280, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.093028] env[62814]: DEBUG nova.compute.manager [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 651.107922] env[62814]: DEBUG oslo_vmware.api [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293279, 'name': PowerOffVM_Task, 'duration_secs': 0.224911} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.111406] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 651.111608] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 651.111977] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e7fda585-3399-4a4a-837a-94302f53e412 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.176671] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 651.176671] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 651.176671] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Deleting the datastore file [datastore2] f5504590-823d-4c37-8226-26684f7957a6 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 651.179723] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18751642-d98d-4be4-a63b-f8d50ad3615c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.192719] env[62814]: DEBUG oslo_vmware.api [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for the task: (returnval){ [ 651.192719] env[62814]: value = "task-4293282" [ 651.192719] env[62814]: _type = "Task" [ 651.192719] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.203583] env[62814]: DEBUG oslo_vmware.api [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293282, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.220378] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Releasing lock "refresh_cache-591e4a92-7c03-4d70-885f-49a66032c97e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 651.220846] env[62814]: DEBUG nova.compute.manager [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 651.221113] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 651.224782] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18c0129-2e4d-4d54-bb82-39473e2da4b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.232982] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 651.234166] env[62814]: DEBUG nova.network.neutron [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Updated VIF entry in instance network info cache for port 80990e09-7766-4c2a-937b-52c703caa252. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 651.234522] env[62814]: DEBUG nova.network.neutron [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Updating instance_info_cache with network_info: [{"id": "80990e09-7766-4c2a-937b-52c703caa252", "address": "fa:16:3e:92:a4:40", "network": {"id": "b4436b99-a4a3-46c9-918a-4e6f6f6bb55e", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-2020231604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d51711cd907404b8d0a48d904b599c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98d96b75-ac36-499a-adc2-130c8c1d55ca", "external-id": "nsx-vlan-transportzone-564", "segmentation_id": 564, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80990e09-77", "ovs_interfaceid": "80990e09-7766-4c2a-937b-52c703caa252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.235687] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-287a1bfd-7e17-4985-96b3-bf1f57718eb7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.242359] env[62814]: DEBUG oslo_vmware.api [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 651.242359] env[62814]: value = "task-4293284" [ 651.242359] env[62814]: _type = "Task" [ 651.242359] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.255366] env[62814]: DEBUG oslo_vmware.api [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293284, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.420060] env[62814]: DEBUG nova.network.neutron [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.434633] env[62814]: DEBUG nova.compute.manager [req-2bb6f98c-ce2e-4b67-aa26-a0c081930147 req-8d5201fb-6471-4567-ae3d-3694ced9a934 service nova] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Received event network-vif-deleted-cdb5c7dc-7e71-4338-af80-0211db859c91 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 651.481556] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293280, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.636528] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b80dafb-93f7-472a-84b8-5990bde0ac77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.646277] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "59d21ef7-df97-49ac-9329-4c18df6dd087" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.646525] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 651.651242] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97705351-c7a1-47d4-a7b9-1a3f289cc0bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.695698] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc866bf-a2e8-48fa-b6e6-4d1d6d448c0a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.711073] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176b7216-6ba2-4e60-b6c8-8cd735fb8251 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.715391] env[62814]: DEBUG oslo_vmware.api [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Task: {'id': task-4293282, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237204} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.716486] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 651.716725] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 651.716912] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 651.717118] env[62814]: INFO nova.compute.manager [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] [instance: f5504590-823d-4c37-8226-26684f7957a6] Took 1.16 seconds to destroy the instance on the hypervisor. [ 651.717386] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 651.718365] env[62814]: DEBUG nova.compute.manager [-] [instance: f5504590-823d-4c37-8226-26684f7957a6] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 651.718365] env[62814]: DEBUG nova.network.neutron [-] [instance: f5504590-823d-4c37-8226-26684f7957a6] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 651.729021] env[62814]: DEBUG nova.compute.provider_tree [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.739181] env[62814]: DEBUG oslo_concurrency.lockutils [req-e73b21b5-85fb-4b41-8788-137020b10418 req-ff6ac7d6-08ca-4350-bfeb-55f982a0374e service nova] Releasing lock "refresh_cache-5087b202-9bba-4489-823b-5d93cbf116e2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 651.758104] env[62814]: DEBUG oslo_vmware.api [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293284, 'name': PowerOffVM_Task, 'duration_secs': 0.226031} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.758104] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 651.758550] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 651.758610] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfb0b792-8093-4f30-9388-46100856e0d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.787167] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 651.787799] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 651.787799] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Deleting the datastore file [datastore2] 591e4a92-7c03-4d70-885f-49a66032c97e {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 651.789034] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca3e5618-23bc-41b6-86f7-1fa56e5195da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.794852] env[62814]: DEBUG oslo_vmware.api [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for the task: (returnval){ [ 651.794852] env[62814]: value = "task-4293286" [ 651.794852] env[62814]: _type = "Task" [ 651.794852] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.803355] env[62814]: DEBUG oslo_vmware.api [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293286, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.838988] env[62814]: DEBUG nova.network.neutron [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Updating instance_info_cache with network_info: [{"id": "22bf5740-3f2e-45ed-be98-ac62c70cfb3b", "address": "fa:16:3e:6e:e4:ef", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22bf5740-3f", "ovs_interfaceid": "22bf5740-3f2e-45ed-be98-ac62c70cfb3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.977076] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293280, 'name': PowerOnVM_Task} progress is 82%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.019215] env[62814]: DEBUG nova.network.neutron [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Successfully created port: 25fab244-2cae-416b-86e1-f25d1ea5aae7 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.114489] env[62814]: DEBUG nova.compute.manager [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 652.141199] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 652.141449] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.141605] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 652.141782] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.141928] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 652.142137] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 652.142363] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 652.142525] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 652.142698] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 652.142853] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 652.143039] env[62814]: DEBUG nova.virt.hardware [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 652.143904] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9480a1e3-1cad-4174-bdba-10ba527c06c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.151468] env[62814]: DEBUG nova.compute.manager [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 652.155211] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47886f14-15c0-4ef9-8f68-85d0d798dff8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.234593] env[62814]: DEBUG nova.scheduler.client.report [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 652.309064] env[62814]: DEBUG oslo_vmware.api [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Task: {'id': task-4293286, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204796} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.309533] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 652.309844] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 652.310223] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.310659] env[62814]: INFO nova.compute.manager [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Took 1.09 seconds to destroy the instance on the hypervisor. [ 652.311040] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 652.313018] env[62814]: DEBUG nova.compute.manager [-] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 652.313018] env[62814]: DEBUG nova.network.neutron [-] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 652.345024] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 652.345024] env[62814]: DEBUG nova.compute.manager [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Instance network_info: |[{"id": "22bf5740-3f2e-45ed-be98-ac62c70cfb3b", "address": "fa:16:3e:6e:e4:ef", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22bf5740-3f", "ovs_interfaceid": "22bf5740-3f2e-45ed-be98-ac62c70cfb3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 652.345918] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:e4:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c621a9c-66f5-426a-8aab-bd8b2e912106', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '22bf5740-3f2e-45ed-be98-ac62c70cfb3b', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 652.351084] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Creating folder: Project (42a09e557c794726aa4895c1563ea62f). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 652.353828] env[62814]: DEBUG nova.network.neutron [-] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.354299] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-270eeb12-d91e-492a-bf60-6746d7b881da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.371179] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Created folder: Project (42a09e557c794726aa4895c1563ea62f) in parent group-v845547. [ 652.371393] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Creating folder: Instances. Parent ref: group-v845584. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 652.371632] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-20380e7e-4593-43fb-b0f8-2f51afeef342 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.475943] env[62814]: DEBUG oslo_vmware.api [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293280, 'name': PowerOnVM_Task, 'duration_secs': 1.137931} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.477271] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 652.477581] env[62814]: INFO nova.compute.manager [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Took 12.95 seconds to spawn the instance on the hypervisor. [ 652.477768] env[62814]: DEBUG nova.compute.manager [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 652.478056] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Created folder: Instances in parent group-v845584. [ 652.478501] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 652.479205] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe6ad41-9877-4746-9b7c-6873885827cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.481677] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 652.481883] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-971a4397-50c6-442c-8e50-bbd3213a6b93 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.504156] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 652.504156] env[62814]: value = "task-4293289" [ 652.504156] env[62814]: _type = "Task" [ 652.504156] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.515213] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293289, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.690953] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 652.741035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.666s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 652.741035] env[62814]: DEBUG nova.compute.manager [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 652.744245] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.826s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 652.746095] env[62814]: INFO nova.compute.claims [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.858087] env[62814]: DEBUG nova.network.neutron [-] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.900586] env[62814]: DEBUG nova.network.neutron [-] [instance: f5504590-823d-4c37-8226-26684f7957a6] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.019486] env[62814]: INFO nova.compute.manager [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Took 24.01 seconds to build instance. [ 653.024677] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293289, 'name': CreateVM_Task, 'duration_secs': 0.471371} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.024677] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 653.025179] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.025352] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 653.025723] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 653.025984] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c61c367-508c-48b9-a0e8-78df431b8030 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.032053] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 653.032053] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5298ad4c-486d-7df2-f3a4-fa2bf3d26b95" [ 653.032053] env[62814]: _type = "Task" [ 653.032053] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.041470] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5298ad4c-486d-7df2-f3a4-fa2bf3d26b95, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.252116] env[62814]: DEBUG nova.compute.utils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 653.256175] env[62814]: DEBUG nova.compute.manager [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 653.256534] env[62814]: DEBUG nova.network.neutron [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 653.363751] env[62814]: INFO nova.compute.manager [-] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Took 1.05 seconds to deallocate network for instance. [ 653.370406] env[62814]: DEBUG nova.policy [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd7a8a9e6751d4c66ab2f80583343de47', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'df24d7d0c7c240ea960e83b4b809a6b9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 653.403417] env[62814]: INFO nova.compute.manager [-] [instance: f5504590-823d-4c37-8226-26684f7957a6] Took 1.69 seconds to deallocate network for instance. [ 653.528034] env[62814]: DEBUG oslo_concurrency.lockutils [None req-88938ac9-4663-477d-8a16-ed544bc5745e tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "5087b202-9bba-4489-823b-5d93cbf116e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.898s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.542267] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5298ad4c-486d-7df2-f3a4-fa2bf3d26b95, 'name': SearchDatastore_Task, 'duration_secs': 0.097946} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.542868] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.542868] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 653.543087] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.543311] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 653.543377] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 653.543969] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a52f8e84-b6ee-4237-96c2-f1ea1301ed0a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.555639] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 653.555835] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 653.556858] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42e17c7d-eceb-41fe-85f7-762d1d90ca96 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.563212] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 653.563212] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e0894d-752c-28aa-037b-f1773df229d2" [ 653.563212] env[62814]: _type = "Task" [ 653.563212] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.574815] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e0894d-752c-28aa-037b-f1773df229d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.756909] env[62814]: DEBUG nova.compute.manager [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 653.869266] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.911054] env[62814]: DEBUG oslo_concurrency.lockutils [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 654.078690] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e0894d-752c-28aa-037b-f1773df229d2, 'name': SearchDatastore_Task, 'duration_secs': 0.014318} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.081412] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-278ec3c0-bf24-46ff-becc-f1586e8968eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.087648] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 654.087648] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52722bdb-8580-9970-45cb-2cd1647f55ee" [ 654.087648] env[62814]: _type = "Task" [ 654.087648] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.097382] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52722bdb-8580-9970-45cb-2cd1647f55ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.220978] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb92dbe-4363-4326-ae27-f4fe99bef151 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.230983] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f76ab6-ae6f-4388-8c01-ccdc304b4106 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.277670] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a615bbe-c10b-4511-a7e0-1b1ab6539a79 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.281314] env[62814]: DEBUG nova.compute.manager [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Received event network-vif-plugged-22bf5740-3f2e-45ed-be98-ac62c70cfb3b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 654.281517] env[62814]: DEBUG oslo_concurrency.lockutils [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] Acquiring lock "c7cef7f3-11db-44e1-a454-98830b465b52-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 654.281719] env[62814]: DEBUG oslo_concurrency.lockutils [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] Lock "c7cef7f3-11db-44e1-a454-98830b465b52-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 654.281886] env[62814]: DEBUG oslo_concurrency.lockutils [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] Lock "c7cef7f3-11db-44e1-a454-98830b465b52-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 654.283652] env[62814]: DEBUG nova.compute.manager [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] No waiting events found dispatching network-vif-plugged-22bf5740-3f2e-45ed-be98-ac62c70cfb3b {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 654.284041] env[62814]: WARNING nova.compute.manager [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Received unexpected event network-vif-plugged-22bf5740-3f2e-45ed-be98-ac62c70cfb3b for instance with vm_state building and task_state spawning. [ 654.284243] env[62814]: DEBUG nova.compute.manager [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Received event network-changed-22bf5740-3f2e-45ed-be98-ac62c70cfb3b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 654.284404] env[62814]: DEBUG nova.compute.manager [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Refreshing instance network info cache due to event network-changed-22bf5740-3f2e-45ed-be98-ac62c70cfb3b. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 654.284602] env[62814]: DEBUG oslo_concurrency.lockutils [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] Acquiring lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.288233] env[62814]: DEBUG oslo_concurrency.lockutils [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] Acquired lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 654.288488] env[62814]: DEBUG nova.network.neutron [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Refreshing network info cache for port 22bf5740-3f2e-45ed-be98-ac62c70cfb3b {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 654.299081] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f8c76d-ba27-4724-8921-2b80f10ca863 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.329285] env[62814]: DEBUG nova.compute.provider_tree [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.398390] env[62814]: DEBUG nova.network.neutron [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Successfully created port: 4615b448-a51a-4bfb-85fb-fd8f118e479d {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.602597] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52722bdb-8580-9970-45cb-2cd1647f55ee, 'name': SearchDatastore_Task, 'duration_secs': 0.024839} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.603047] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 654.603443] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c7cef7f3-11db-44e1-a454-98830b465b52/c7cef7f3-11db-44e1-a454-98830b465b52.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 654.603823] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a8a1226-a3aa-4ecc-a41a-d1418e31710b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.613045] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 654.613045] env[62814]: value = "task-4293291" [ 654.613045] env[62814]: _type = "Task" [ 654.613045] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.625245] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293291, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.757659] env[62814]: DEBUG nova.network.neutron [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Successfully updated port: 25fab244-2cae-416b-86e1-f25d1ea5aae7 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 654.777650] env[62814]: DEBUG nova.compute.manager [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 654.814519] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 654.814863] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 654.814987] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 654.815218] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 654.815371] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 654.815520] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 654.815740] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 654.816465] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 654.816726] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 654.816907] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 654.817144] env[62814]: DEBUG nova.virt.hardware [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 654.818101] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2e5655-4561-4ebc-a28f-ff91e7574eb2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.830796] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cce263b5-5a6c-471b-8434-97efd8485bfa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.840251] env[62814]: DEBUG nova.scheduler.client.report [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 654.935338] env[62814]: DEBUG nova.compute.manager [req-2e61738f-12dc-4925-846f-b072bd836ce9 req-14c422d8-b945-4f02-88f9-35c6acabbd06 service nova] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Received event network-vif-deleted-320f12a1-7803-4527-847e-8660c500ed93 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 654.959069] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquiring lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 654.959361] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.032135] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "62bc755d-4f96-4486-884b-0d0c337267aa" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.032358] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "62bc755d-4f96-4486-884b-0d0c337267aa" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.032903] env[62814]: DEBUG nova.compute.manager [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 655.033824] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e1cb07-5f89-4b59-8402-549ee1d306ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.043531] env[62814]: DEBUG nova.compute.manager [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 655.044581] env[62814]: DEBUG nova.objects.instance [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lazy-loading 'flavor' on Instance uuid 62bc755d-4f96-4486-884b-0d0c337267aa {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 655.126052] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293291, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.185795] env[62814]: DEBUG nova.network.neutron [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Updated VIF entry in instance network info cache for port 22bf5740-3f2e-45ed-be98-ac62c70cfb3b. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 655.186234] env[62814]: DEBUG nova.network.neutron [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Updating instance_info_cache with network_info: [{"id": "22bf5740-3f2e-45ed-be98-ac62c70cfb3b", "address": "fa:16:3e:6e:e4:ef", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22bf5740-3f", "ovs_interfaceid": "22bf5740-3f2e-45ed-be98-ac62c70cfb3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.260576] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquiring lock "refresh_cache-3d3f32fc-276a-49be-b471-01a5d6fc5069" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.260755] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquired lock "refresh_cache-3d3f32fc-276a-49be-b471-01a5d6fc5069" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.260903] env[62814]: DEBUG nova.network.neutron [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.347370] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.603s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.348056] env[62814]: DEBUG nova.compute.manager [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 655.351356] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.490s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.351591] env[62814]: DEBUG nova.objects.instance [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lazy-loading 'resources' on Instance uuid 37649169-66e6-4ef1-a4cd-c65b9831004a {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 655.462208] env[62814]: DEBUG nova.compute.manager [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 655.626309] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293291, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585322} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.626681] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c7cef7f3-11db-44e1-a454-98830b465b52/c7cef7f3-11db-44e1-a454-98830b465b52.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 655.626955] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 655.627232] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c459f503-d03f-4993-9574-70a02d23b210 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.636654] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 655.636654] env[62814]: value = "task-4293293" [ 655.636654] env[62814]: _type = "Task" [ 655.636654] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.645574] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293293, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.690243] env[62814]: DEBUG oslo_concurrency.lockutils [req-ba61ddf9-7b3b-4093-b58b-49a2106cc922 req-dafa8730-ffca-410c-bb3b-59829ce2c00c service nova] Releasing lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 655.842077] env[62814]: DEBUG nova.network.neutron [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.858171] env[62814]: DEBUG nova.compute.utils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 655.863374] env[62814]: DEBUG nova.compute.manager [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Not allocating networking since 'none' was specified. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 655.996529] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 656.059696] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 656.061273] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7876899-21d3-49ac-ac93-c7220a6f08b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.069254] env[62814]: DEBUG oslo_vmware.api [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 656.069254] env[62814]: value = "task-4293294" [ 656.069254] env[62814]: _type = "Task" [ 656.069254] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.083807] env[62814]: DEBUG oslo_vmware.api [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293294, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.148448] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293293, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118696} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.148776] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 656.149537] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72837bc4-2269-4943-b819-a2d518d956e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.152751] env[62814]: DEBUG nova.network.neutron [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Updating instance_info_cache with network_info: [{"id": "25fab244-2cae-416b-86e1-f25d1ea5aae7", "address": "fa:16:3e:b8:89:0d", "network": {"id": "8a6a9999-88fc-4c83-b519-b2b9a657ab49", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1596604004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07ee864ce77f44de84624ee579fb5316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25fab244-2c", "ovs_interfaceid": "25fab244-2cae-416b-86e1-f25d1ea5aae7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.176800] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Reconfiguring VM instance instance-0000000c to attach disk [datastore2] c7cef7f3-11db-44e1-a454-98830b465b52/c7cef7f3-11db-44e1-a454-98830b465b52.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 656.183021] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec3fc228-60b4-462a-a3fb-ce08b142f84c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.201615] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 656.201615] env[62814]: value = "task-4293295" [ 656.201615] env[62814]: _type = "Task" [ 656.201615] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.211988] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293295, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.301857] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd7cff9-fb7a-4d18-9b9c-a7d9e3cf0ecd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.311894] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06eaac48-4be0-4afd-b2a8-a289e42fafec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.347594] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fa7316-3349-4e43-8e1f-8b426a8b1ad4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.356429] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af85071-99b4-411f-a10e-be5086338704 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.370932] env[62814]: DEBUG nova.compute.manager [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 656.374539] env[62814]: DEBUG nova.compute.provider_tree [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.590743] env[62814]: DEBUG oslo_vmware.api [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293294, 'name': PowerOffVM_Task, 'duration_secs': 0.331849} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.590743] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 656.590743] env[62814]: DEBUG nova.compute.manager [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 656.590743] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d6a1c6a-c9e0-483b-bfad-d5b8273891b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.656050] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Releasing lock "refresh_cache-3d3f32fc-276a-49be-b471-01a5d6fc5069" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 656.656601] env[62814]: DEBUG nova.compute.manager [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Instance network_info: |[{"id": "25fab244-2cae-416b-86e1-f25d1ea5aae7", "address": "fa:16:3e:b8:89:0d", "network": {"id": "8a6a9999-88fc-4c83-b519-b2b9a657ab49", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1596604004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07ee864ce77f44de84624ee579fb5316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25fab244-2c", "ovs_interfaceid": "25fab244-2cae-416b-86e1-f25d1ea5aae7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 656.656886] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b8:89:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78340140-126f-4ef8-a340-debaa64da3e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25fab244-2cae-416b-86e1-f25d1ea5aae7', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 656.671621] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Creating folder: Project (07ee864ce77f44de84624ee579fb5316). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 656.673369] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8383d80b-68b6-4ffa-8009-847686a43aea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.684579] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Created folder: Project (07ee864ce77f44de84624ee579fb5316) in parent group-v845547. [ 656.684579] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Creating folder: Instances. Parent ref: group-v845588. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 656.690956] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-45e4fb59-4252-483a-bc10-31460c324ca8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.703273] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Created folder: Instances in parent group-v845588. [ 656.703273] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 656.703273] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 656.708184] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf621f43-dca7-45f5-a2d3-8c18ad849730 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.739162] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293295, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.742996] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 656.742996] env[62814]: value = "task-4293298" [ 656.742996] env[62814]: _type = "Task" [ 656.742996] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.752310] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293298, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.879815] env[62814]: DEBUG nova.scheduler.client.report [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 656.995570] env[62814]: DEBUG nova.network.neutron [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Successfully updated port: 4615b448-a51a-4bfb-85fb-fd8f118e479d {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 657.102229] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44d29f05-c066-4ddf-b69e-ccf24773af18 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "62bc755d-4f96-4486-884b-0d0c337267aa" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.069s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.155313] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "4ed66b36-b6c6-4673-9c03-169a01134574" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 657.155313] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "4ed66b36-b6c6-4673-9c03-169a01134574" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.212991] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293295, 'name': ReconfigVM_Task, 'duration_secs': 0.582823} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.213694] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Reconfigured VM instance instance-0000000c to attach disk [datastore2] c7cef7f3-11db-44e1-a454-98830b465b52/c7cef7f3-11db-44e1-a454-98830b465b52.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 657.215743] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4c9db3d1-4648-42a0-8aab-9d575bd1953d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.225017] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 657.225017] env[62814]: value = "task-4293299" [ 657.225017] env[62814]: _type = "Task" [ 657.225017] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.231883] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293299, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.250979] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293298, 'name': CreateVM_Task, 'duration_secs': 0.349752} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.251152] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 657.251809] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.251970] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.252313] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 657.252903] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b264ca3e-f390-408c-b37c-b35fac521759 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.257514] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for the task: (returnval){ [ 657.257514] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b040a5-5395-218e-820c-500dd629ac4d" [ 657.257514] env[62814]: _type = "Task" [ 657.257514] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.265843] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b040a5-5395-218e-820c-500dd629ac4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.387744] env[62814]: DEBUG nova.compute.manager [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 657.390912] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.039s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.393044] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.547s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.395358] env[62814]: INFO nova.compute.claims [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.436339] env[62814]: INFO nova.scheduler.client.report [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Deleted allocations for instance 37649169-66e6-4ef1-a4cd-c65b9831004a [ 657.445283] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 657.445714] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.445797] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 657.445911] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.446063] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 657.446214] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 657.446442] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 657.446573] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 657.446736] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 657.446894] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 657.447074] env[62814]: DEBUG nova.virt.hardware [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 657.448310] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d797c88b-e8ae-44a6-826a-5d1dabb5edc5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.465650] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4591be1d-733f-498a-8a85-b830f6458a7e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.488417] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 657.498675] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Creating folder: Project (b938974f61fc4fa3b50cce1a126a9ef7). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 657.498675] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3201b204-0bae-455f-be68-d96e0dff0e64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.500393] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquiring lock "refresh_cache-6976b964-a8d3-4886-8aac-8d513e721018" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.500393] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquired lock "refresh_cache-6976b964-a8d3-4886-8aac-8d513e721018" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.500393] env[62814]: DEBUG nova.network.neutron [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.509718] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Created folder: Project (b938974f61fc4fa3b50cce1a126a9ef7) in parent group-v845547. [ 657.509989] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Creating folder: Instances. Parent ref: group-v845591. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 657.510261] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e07b910-1017-4c53-8e68-02507ab91d23 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.524114] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Created folder: Instances in parent group-v845591. [ 657.524310] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 657.524512] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 657.524706] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-711e4f83-5561-408f-a24d-8b020a55b427 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.549536] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 657.549536] env[62814]: value = "task-4293302" [ 657.549536] env[62814]: _type = "Task" [ 657.549536] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.563591] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293302, 'name': CreateVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.733700] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293299, 'name': Rename_Task, 'duration_secs': 0.251513} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.733916] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 657.734465] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80b0e20b-28b8-4665-8cce-04f87574c8c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.741344] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 657.741344] env[62814]: value = "task-4293303" [ 657.741344] env[62814]: _type = "Task" [ 657.741344] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.752710] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293303, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.769066] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b040a5-5395-218e-820c-500dd629ac4d, 'name': SearchDatastore_Task, 'duration_secs': 0.008911} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.769066] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 657.769383] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 657.769575] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.769636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.770012] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 657.771408] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65671592-89f4-446e-9ade-79c8a3e9ff6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.784335] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 657.784335] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 657.784335] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25061b5d-f8b9-471c-ac4b-c3396844ea7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.790916] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for the task: (returnval){ [ 657.790916] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5292772d-3ffc-d494-451e-680c81a11a4c" [ 657.790916] env[62814]: _type = "Task" [ 657.790916] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.799621] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5292772d-3ffc-d494-451e-680c81a11a4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.960213] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a3653573-1e86-4769-aa0a-6de6827fed7f tempest-ImagesNegativeTestJSON-1161638411 tempest-ImagesNegativeTestJSON-1161638411-project-member] Lock "37649169-66e6-4ef1-a4cd-c65b9831004a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.110s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 658.061724] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293302, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.062608] env[62814]: DEBUG nova.network.neutron [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.252291] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293303, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.292817] env[62814]: DEBUG nova.compute.manager [req-54c2f614-8ab2-4cbb-9db3-7b507daca5fd req-16121e45-dd04-4056-850c-0ad0ed5f5cef service nova] [instance: f5504590-823d-4c37-8226-26684f7957a6] Received event network-vif-deleted-e4cf948d-498b-4313-851a-72093b0a656c {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 658.303887] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5292772d-3ffc-d494-451e-680c81a11a4c, 'name': SearchDatastore_Task, 'duration_secs': 0.010724} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.305183] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06bcd82d-4b45-46d3-a190-7e52421821d6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.316948] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for the task: (returnval){ [ 658.316948] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5251b7f4-c7a1-baaf-ab88-84e3d1ca5dbd" [ 658.316948] env[62814]: _type = "Task" [ 658.316948] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.325154] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5251b7f4-c7a1-baaf-ab88-84e3d1ca5dbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.481409] env[62814]: DEBUG nova.network.neutron [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Updating instance_info_cache with network_info: [{"id": "4615b448-a51a-4bfb-85fb-fd8f118e479d", "address": "fa:16:3e:97:b0:71", "network": {"id": "50951181-5761-45cf-9bd9-9f6766592c6b", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-505857443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df24d7d0c7c240ea960e83b4b809a6b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b911797-478d-4ee5-bce9-6f2f49014e94", "external-id": "nsx-vlan-transportzone-70", "segmentation_id": 70, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4615b448-a5", "ovs_interfaceid": "4615b448-a51a-4bfb-85fb-fd8f118e479d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.566613] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293302, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.703333] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "29e69c7c-08f7-4da4-9509-02a94bf971bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 658.703997] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "29e69c7c-08f7-4da4-9509-02a94bf971bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 658.752392] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293303, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.831807] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5251b7f4-c7a1-baaf-ab88-84e3d1ca5dbd, 'name': SearchDatastore_Task, 'duration_secs': 0.010771} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.831807] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 658.831807] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 3d3f32fc-276a-49be-b471-01a5d6fc5069/3d3f32fc-276a-49be-b471-01a5d6fc5069.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 658.836163] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6d5251b-af73-42d0-9120-6a75e1fd8f29 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.846822] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for the task: (returnval){ [ 658.846822] env[62814]: value = "task-4293304" [ 658.846822] env[62814]: _type = "Task" [ 658.846822] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.855388] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293304, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.871417] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e0fed4-083b-488e-a191-2f4742521e5f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.879343] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106236bb-a3ff-472b-b88b-5dd72f5723c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.921387] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a96f39f-c7b1-41f4-81ec-7525497293e1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.935130] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745eaff0-1b09-40b3-a0d2-9181aea3ef19 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.958025] env[62814]: DEBUG nova.compute.provider_tree [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.969200] env[62814]: DEBUG nova.compute.manager [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Received event network-changed-6e1ae96b-a9fd-4e11-9803-191a35a77cb2 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 658.969431] env[62814]: DEBUG nova.compute.manager [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Refreshing instance network info cache due to event network-changed-6e1ae96b-a9fd-4e11-9803-191a35a77cb2. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 658.969657] env[62814]: DEBUG oslo_concurrency.lockutils [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] Acquiring lock "refresh_cache-17bfe703-ff96-4cb7-8535-eed57e10d673" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.969803] env[62814]: DEBUG oslo_concurrency.lockutils [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] Acquired lock "refresh_cache-17bfe703-ff96-4cb7-8535-eed57e10d673" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 658.969958] env[62814]: DEBUG nova.network.neutron [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Refreshing network info cache for port 6e1ae96b-a9fd-4e11-9803-191a35a77cb2 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 658.984801] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Releasing lock "refresh_cache-6976b964-a8d3-4886-8aac-8d513e721018" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 658.985243] env[62814]: DEBUG nova.compute.manager [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Instance network_info: |[{"id": "4615b448-a51a-4bfb-85fb-fd8f118e479d", "address": "fa:16:3e:97:b0:71", "network": {"id": "50951181-5761-45cf-9bd9-9f6766592c6b", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-505857443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df24d7d0c7c240ea960e83b4b809a6b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b911797-478d-4ee5-bce9-6f2f49014e94", "external-id": "nsx-vlan-transportzone-70", "segmentation_id": 70, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4615b448-a5", "ovs_interfaceid": "4615b448-a51a-4bfb-85fb-fd8f118e479d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 658.986234] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:b0:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9b911797-478d-4ee5-bce9-6f2f49014e94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4615b448-a51a-4bfb-85fb-fd8f118e479d', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 658.994379] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Creating folder: Project (df24d7d0c7c240ea960e83b4b809a6b9). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 658.998049] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c721b50-3529-4f37-b6b3-792bec5902fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.006011] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Created folder: Project (df24d7d0c7c240ea960e83b4b809a6b9) in parent group-v845547. [ 659.006364] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Creating folder: Instances. Parent ref: group-v845594. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 659.006571] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f5cb1a00-08e5-42e3-8d94-68d7a364bcfb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.016137] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Created folder: Instances in parent group-v845594. [ 659.016137] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 659.016361] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 659.016440] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a29c253e-1fb5-4a2e-a7a3-5ecba0e39338 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.038172] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 659.038172] env[62814]: value = "task-4293307" [ 659.038172] env[62814]: _type = "Task" [ 659.038172] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.046522] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293307, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.062329] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293302, 'name': CreateVM_Task, 'duration_secs': 1.286899} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.062540] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 659.062989] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.063171] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.063516] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 659.063794] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c488767-42e5-4f7e-b180-fe489737e69c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.070835] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 659.070835] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c215d4-fc90-a429-0358-5cfe154eb8ba" [ 659.070835] env[62814]: _type = "Task" [ 659.070835] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.080781] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c215d4-fc90-a429-0358-5cfe154eb8ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.253165] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293303, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.357738] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293304, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502777} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.357738] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 3d3f32fc-276a-49be-b471-01a5d6fc5069/3d3f32fc-276a-49be-b471-01a5d6fc5069.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 659.357928] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 659.358180] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1a0fadb2-9163-43e4-a455-49a40aa398ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.365538] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for the task: (returnval){ [ 659.365538] env[62814]: value = "task-4293308" [ 659.365538] env[62814]: _type = "Task" [ 659.365538] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.373570] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293308, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.458289] env[62814]: DEBUG nova.scheduler.client.report [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 659.548991] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293307, 'name': CreateVM_Task, 'duration_secs': 0.506592} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.551375] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 659.554693] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.582639] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c215d4-fc90-a429-0358-5cfe154eb8ba, 'name': SearchDatastore_Task, 'duration_secs': 0.062907} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.583214] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.584039] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 659.584473] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.584570] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.584803] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 659.585201] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 659.585661] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 659.585950] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b76f31b-6393-4466-bf8e-8eb89c934ea2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.588676] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccf51f28-882b-4311-9016-d26a9c8fa2f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.595993] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for the task: (returnval){ [ 659.595993] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526285ca-e96d-ac39-d393-cb2a746f9f57" [ 659.595993] env[62814]: _type = "Task" [ 659.595993] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.597365] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 659.597539] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 659.602458] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1cd4aaa-6bd1-4b4e-9d8a-3d3583be1e08 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.612189] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 659.612189] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 659.621200] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526285ca-e96d-ac39-d393-cb2a746f9f57, 'name': SearchDatastore_Task, 'duration_secs': 0.010843} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.621200] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 659.621200] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52147e61-565a-6015-2823-f89825396cc2" [ 659.621200] env[62814]: _type = "Task" [ 659.621200] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.622102] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.622523] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 659.622896] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.634162] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52147e61-565a-6015-2823-f89825396cc2, 'name': SearchDatastore_Task, 'duration_secs': 0.009392} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.634963] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68059e4a-91f0-41a4-86c7-5276545a0ed8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.640436] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 659.640436] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a932da-6a3a-57a9-0596-9e6e25e9dd9f" [ 659.640436] env[62814]: _type = "Task" [ 659.640436] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.649925] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a932da-6a3a-57a9-0596-9e6e25e9dd9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.759459] env[62814]: DEBUG oslo_vmware.api [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293303, 'name': PowerOnVM_Task, 'duration_secs': 1.977354} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.759808] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 659.760047] env[62814]: INFO nova.compute.manager [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Took 11.40 seconds to spawn the instance on the hypervisor. [ 659.760226] env[62814]: DEBUG nova.compute.manager [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 659.761072] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95aa8209-5d14-42fe-8f6a-6fafe1be86ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.875449] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293308, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104672} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.875449] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 659.876452] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50e4696-4301-4152-b92b-048a26e5aff4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.904100] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Reconfiguring VM instance instance-0000000d to attach disk [datastore2] 3d3f32fc-276a-49be-b471-01a5d6fc5069/3d3f32fc-276a-49be-b471-01a5d6fc5069.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 659.905893] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cf6c4ad8-03ec-41a7-8305-d08e3abf37be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.927743] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for the task: (returnval){ [ 659.927743] env[62814]: value = "task-4293309" [ 659.927743] env[62814]: _type = "Task" [ 659.927743] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.936359] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293309, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.965474] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.572s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.966059] env[62814]: DEBUG nova.compute.manager [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 659.970175] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.619s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 659.970175] env[62814]: DEBUG nova.objects.instance [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lazy-loading 'resources' on Instance uuid 32f1bd48-3a2e-42fe-8033-85d3ad55affe {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 659.990808] env[62814]: DEBUG nova.network.neutron [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Updated VIF entry in instance network info cache for port 6e1ae96b-a9fd-4e11-9803-191a35a77cb2. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 659.991380] env[62814]: DEBUG nova.network.neutron [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Updating instance_info_cache with network_info: [{"id": "6e1ae96b-a9fd-4e11-9803-191a35a77cb2", "address": "fa:16:3e:3c:b9:dc", "network": {"id": "b16d5204-f140-44fb-b4c4-621c3f5a3423", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-309109771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5f0190bd5f24a909c9839b7fd1e870b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec46b14d-3310-4f2b-96c1-f53ee47d3759", "external-id": "nsx-vlan-transportzone-323", "segmentation_id": 323, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e1ae96b-a9", "ovs_interfaceid": "6e1ae96b-a9fd-4e11-9803-191a35a77cb2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.152973] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a932da-6a3a-57a9-0596-9e6e25e9dd9f, 'name': SearchDatastore_Task, 'duration_secs': 0.009447} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.153519] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 660.153976] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 296f6c11-7108-42e6-8ada-5d8c08b00da6/296f6c11-7108-42e6-8ada-5d8c08b00da6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 660.154431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.154740] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 660.155099] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-01096da9-a4bf-4aa2-bb9b-146a1f98834a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.158343] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4baa320-46b3-413c-bf8e-8db552ba6f1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.167809] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 660.167809] env[62814]: value = "task-4293310" [ 660.167809] env[62814]: _type = "Task" [ 660.167809] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.171208] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 660.171348] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 660.173283] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d8d8b0b-e4db-4b7f-9938-ba125d2f672e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.178207] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293310, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.181538] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for the task: (returnval){ [ 660.181538] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f52c72-c983-ee67-8ba7-810339b50908" [ 660.181538] env[62814]: _type = "Task" [ 660.181538] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.190141] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f52c72-c983-ee67-8ba7-810339b50908, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.281581] env[62814]: INFO nova.compute.manager [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Took 28.35 seconds to build instance. [ 660.440881] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293309, 'name': ReconfigVM_Task, 'duration_secs': 0.292546} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.441095] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Reconfigured VM instance instance-0000000d to attach disk [datastore2] 3d3f32fc-276a-49be-b471-01a5d6fc5069/3d3f32fc-276a-49be-b471-01a5d6fc5069.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 660.441799] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c84f215c-34e3-4a1c-b60a-89c6d9dd89c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.450525] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for the task: (returnval){ [ 660.450525] env[62814]: value = "task-4293311" [ 660.450525] env[62814]: _type = "Task" [ 660.450525] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.462212] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293311, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.473861] env[62814]: DEBUG nova.compute.utils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 660.478077] env[62814]: DEBUG nova.compute.manager [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 660.478393] env[62814]: DEBUG nova.network.neutron [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 660.494815] env[62814]: DEBUG oslo_concurrency.lockutils [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] Releasing lock "refresh_cache-17bfe703-ff96-4cb7-8535-eed57e10d673" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 660.495111] env[62814]: DEBUG nova.compute.manager [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Received event network-vif-plugged-25fab244-2cae-416b-86e1-f25d1ea5aae7 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 660.495319] env[62814]: DEBUG oslo_concurrency.lockutils [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] Acquiring lock "3d3f32fc-276a-49be-b471-01a5d6fc5069-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 660.495731] env[62814]: DEBUG oslo_concurrency.lockutils [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 660.495731] env[62814]: DEBUG oslo_concurrency.lockutils [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 660.495909] env[62814]: DEBUG nova.compute.manager [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] No waiting events found dispatching network-vif-plugged-25fab244-2cae-416b-86e1-f25d1ea5aae7 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 660.495989] env[62814]: WARNING nova.compute.manager [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Received unexpected event network-vif-plugged-25fab244-2cae-416b-86e1-f25d1ea5aae7 for instance with vm_state building and task_state spawning. [ 660.496178] env[62814]: DEBUG nova.compute.manager [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Received event network-changed-25fab244-2cae-416b-86e1-f25d1ea5aae7 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 660.496360] env[62814]: DEBUG nova.compute.manager [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Refreshing instance network info cache due to event network-changed-25fab244-2cae-416b-86e1-f25d1ea5aae7. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 660.496518] env[62814]: DEBUG oslo_concurrency.lockutils [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] Acquiring lock "refresh_cache-3d3f32fc-276a-49be-b471-01a5d6fc5069" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.496663] env[62814]: DEBUG oslo_concurrency.lockutils [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] Acquired lock "refresh_cache-3d3f32fc-276a-49be-b471-01a5d6fc5069" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.496815] env[62814]: DEBUG nova.network.neutron [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Refreshing network info cache for port 25fab244-2cae-416b-86e1-f25d1ea5aae7 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 660.629776] env[62814]: DEBUG nova.policy [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '307242d782d24d1594ee89c0e46e7aa3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42a09e557c794726aa4895c1563ea62f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 660.678091] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293310, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.696418] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f52c72-c983-ee67-8ba7-810339b50908, 'name': SearchDatastore_Task, 'duration_secs': 0.012176} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.699025] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab131220-dae8-4874-9ec4-a9573dc46cbe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.704167] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for the task: (returnval){ [ 660.704167] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e910ba-6a95-0cf9-5aa4-dd83c9881d97" [ 660.704167] env[62814]: _type = "Task" [ 660.704167] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.712541] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e910ba-6a95-0cf9-5aa4-dd83c9881d97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.784028] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f6b5ec-3512-499b-a5cb-0885332ed9cd tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "c7cef7f3-11db-44e1-a454-98830b465b52" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.869s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 660.962586] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293311, 'name': Rename_Task, 'duration_secs': 0.323614} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.962878] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 660.963524] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-595cf14b-78f1-4512-81db-107962daf7a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.971467] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for the task: (returnval){ [ 660.971467] env[62814]: value = "task-4293312" [ 660.971467] env[62814]: _type = "Task" [ 660.971467] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.980180] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293312, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.982311] env[62814]: DEBUG nova.compute.manager [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 660.992259] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14ebd97-99b2-469c-a41a-d420ae27bc07 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.997155] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c3f2a2-d20b-4626-8cfd-e76d1adc0a97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.033689] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26add10f-b042-47b0-9f32-c383d1b60016 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.041667] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-797cf423-0427-4277-99c8-3bd8e73de6b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.057609] env[62814]: DEBUG nova.compute.provider_tree [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.177580] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293310, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528435} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.177838] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 296f6c11-7108-42e6-8ada-5d8c08b00da6/296f6c11-7108-42e6-8ada-5d8c08b00da6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 661.178064] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 661.178344] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6da914b-757b-47ec-88db-85b48418990a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.187025] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 661.187025] env[62814]: value = "task-4293313" [ 661.187025] env[62814]: _type = "Task" [ 661.187025] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.196319] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293313, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.219450] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e910ba-6a95-0cf9-5aa4-dd83c9881d97, 'name': SearchDatastore_Task, 'duration_secs': 0.010711} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.219822] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 661.220132] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 6976b964-a8d3-4886-8aac-8d513e721018/6976b964-a8d3-4886-8aac-8d513e721018.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 661.220638] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c0115f91-cb41-4ea1-a531-f7b9e749b97f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.229757] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for the task: (returnval){ [ 661.229757] env[62814]: value = "task-4293314" [ 661.229757] env[62814]: _type = "Task" [ 661.229757] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.240249] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293314, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.294763] env[62814]: DEBUG nova.compute.manager [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 661.493644] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293312, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.561292] env[62814]: DEBUG nova.scheduler.client.report [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 661.573989] env[62814]: DEBUG nova.network.neutron [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Successfully created port: a2e379d6-9cd3-4a22-85c2-72628174b4a6 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.700300] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293313, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064475} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.700941] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 661.702401] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405364f7-589d-4944-af41-b06b2267e0a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.733928] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 296f6c11-7108-42e6-8ada-5d8c08b00da6/296f6c11-7108-42e6-8ada-5d8c08b00da6.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 661.733928] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2b13282-2094-4d7b-ba7e-780b54013c6a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.769141] env[62814]: DEBUG nova.network.neutron [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Updated VIF entry in instance network info cache for port 25fab244-2cae-416b-86e1-f25d1ea5aae7. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 661.771322] env[62814]: DEBUG nova.network.neutron [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Updating instance_info_cache with network_info: [{"id": "25fab244-2cae-416b-86e1-f25d1ea5aae7", "address": "fa:16:3e:b8:89:0d", "network": {"id": "8a6a9999-88fc-4c83-b519-b2b9a657ab49", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1596604004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07ee864ce77f44de84624ee579fb5316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25fab244-2c", "ovs_interfaceid": "25fab244-2cae-416b-86e1-f25d1ea5aae7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.786778] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293314, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.786951] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 661.786951] env[62814]: value = "task-4293316" [ 661.786951] env[62814]: _type = "Task" [ 661.786951] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.802265] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293316, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.826961] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 661.984184] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293312, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.998492] env[62814]: DEBUG nova.compute.manager [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 662.041353] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 662.041631] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.041781] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 662.041957] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.043015] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 662.043015] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 662.043015] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 662.043015] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 662.043336] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 662.043336] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 662.043433] env[62814]: DEBUG nova.virt.hardware [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 662.044326] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc1b220-da00-458d-a890-eec7ebb90a86 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.053181] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2767610f-b1d6-46d9-8a32-82f76a346436 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.072346] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.103s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 662.076468] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.170s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 662.081385] env[62814]: INFO nova.compute.claims [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 662.120247] env[62814]: INFO nova.scheduler.client.report [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Deleted allocations for instance 32f1bd48-3a2e-42fe-8033-85d3ad55affe [ 662.246231] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293314, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568015} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.246231] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 6976b964-a8d3-4886-8aac-8d513e721018/6976b964-a8d3-4886-8aac-8d513e721018.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 662.246678] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 662.246678] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc523fdc-dc51-425d-abc9-a86976285b6a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.257013] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for the task: (returnval){ [ 662.257013] env[62814]: value = "task-4293317" [ 662.257013] env[62814]: _type = "Task" [ 662.257013] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.271483] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293317, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.275718] env[62814]: DEBUG oslo_concurrency.lockutils [req-85c13761-4ded-498a-873b-4412d0d25c51 req-a8fe0e67-f3eb-4ba6-a059-03af44ad7ef3 service nova] Releasing lock "refresh_cache-3d3f32fc-276a-49be-b471-01a5d6fc5069" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 662.296854] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293316, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.309378] env[62814]: DEBUG nova.compute.manager [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 662.310353] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b966538c-31cc-45a1-9b09-45a91931fde5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.427932] env[62814]: DEBUG nova.compute.manager [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Received event network-vif-plugged-4615b448-a51a-4bfb-85fb-fd8f118e479d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 662.427932] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] Acquiring lock "6976b964-a8d3-4886-8aac-8d513e721018-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 662.427932] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] Lock "6976b964-a8d3-4886-8aac-8d513e721018-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 662.427932] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] Lock "6976b964-a8d3-4886-8aac-8d513e721018-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 662.427932] env[62814]: DEBUG nova.compute.manager [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] No waiting events found dispatching network-vif-plugged-4615b448-a51a-4bfb-85fb-fd8f118e479d {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 662.428178] env[62814]: WARNING nova.compute.manager [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Received unexpected event network-vif-plugged-4615b448-a51a-4bfb-85fb-fd8f118e479d for instance with vm_state building and task_state spawning. [ 662.428178] env[62814]: DEBUG nova.compute.manager [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Received event network-changed-4615b448-a51a-4bfb-85fb-fd8f118e479d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 662.428178] env[62814]: DEBUG nova.compute.manager [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Refreshing instance network info cache due to event network-changed-4615b448-a51a-4bfb-85fb-fd8f118e479d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 662.428178] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] Acquiring lock "refresh_cache-6976b964-a8d3-4886-8aac-8d513e721018" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.428394] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] Acquired lock "refresh_cache-6976b964-a8d3-4886-8aac-8d513e721018" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 662.428435] env[62814]: DEBUG nova.network.neutron [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Refreshing network info cache for port 4615b448-a51a-4bfb-85fb-fd8f118e479d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 662.488409] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293312, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.638381] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6070028-fe89-452f-a8bc-788ec6c13117 tempest-ServerDiagnosticsNegativeTest-137291582 tempest-ServerDiagnosticsNegativeTest-137291582-project-member] Lock "32f1bd48-3a2e-42fe-8033-85d3ad55affe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.216s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 662.646122] env[62814]: DEBUG nova.compute.manager [req-d3ea7483-2455-4f0c-8879-8c85932ec2ef req-cf1643e4-668c-4bac-961b-8922b65e8b2f service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Received event network-changed-80990e09-7766-4c2a-937b-52c703caa252 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 662.646421] env[62814]: DEBUG nova.compute.manager [req-d3ea7483-2455-4f0c-8879-8c85932ec2ef req-cf1643e4-668c-4bac-961b-8922b65e8b2f service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Refreshing instance network info cache due to event network-changed-80990e09-7766-4c2a-937b-52c703caa252. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 662.646682] env[62814]: DEBUG oslo_concurrency.lockutils [req-d3ea7483-2455-4f0c-8879-8c85932ec2ef req-cf1643e4-668c-4bac-961b-8922b65e8b2f service nova] Acquiring lock "refresh_cache-5087b202-9bba-4489-823b-5d93cbf116e2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.646857] env[62814]: DEBUG oslo_concurrency.lockutils [req-d3ea7483-2455-4f0c-8879-8c85932ec2ef req-cf1643e4-668c-4bac-961b-8922b65e8b2f service nova] Acquired lock "refresh_cache-5087b202-9bba-4489-823b-5d93cbf116e2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 662.647082] env[62814]: DEBUG nova.network.neutron [req-d3ea7483-2455-4f0c-8879-8c85932ec2ef req-cf1643e4-668c-4bac-961b-8922b65e8b2f service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Refreshing network info cache for port 80990e09-7766-4c2a-937b-52c703caa252 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 662.771230] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293317, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068888} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.771230] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 662.771531] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846adce8-d4bc-417b-a679-6e0a97185e3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.795962] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Reconfiguring VM instance instance-0000000e to attach disk [datastore2] 6976b964-a8d3-4886-8aac-8d513e721018/6976b964-a8d3-4886-8aac-8d513e721018.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 662.796672] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-716f2b93-e23a-4e4f-b724-f12310f24677 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.825901] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293316, 'name': ReconfigVM_Task, 'duration_secs': 0.665943} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.828434] env[62814]: INFO nova.compute.manager [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] instance snapshotting [ 662.828686] env[62814]: WARNING nova.compute.manager [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 662.830760] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 296f6c11-7108-42e6-8ada-5d8c08b00da6/296f6c11-7108-42e6-8ada-5d8c08b00da6.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 662.831460] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for the task: (returnval){ [ 662.831460] env[62814]: value = "task-4293318" [ 662.831460] env[62814]: _type = "Task" [ 662.831460] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.831929] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a49a938c-fd74-4a17-80d7-62402431ed34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.834699] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6b764e-68f0-405a-bfcc-f5dcd4768c82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.847825] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293318, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.862687] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 662.862687] env[62814]: value = "task-4293319" [ 662.862687] env[62814]: _type = "Task" [ 662.862687] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.864116] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0e3284-11c9-4c17-ae6b-8727fe8afa98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.881072] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293319, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.988934] env[62814]: DEBUG oslo_vmware.api [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293312, 'name': PowerOnVM_Task, 'duration_secs': 1.696974} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.989781] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 662.989781] env[62814]: INFO nova.compute.manager [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Took 10.87 seconds to spawn the instance on the hypervisor. [ 662.989781] env[62814]: DEBUG nova.compute.manager [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 662.990646] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3845ed55-c7ce-4167-90f1-4ac4c4e48cdf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.346129] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293318, 'name': ReconfigVM_Task, 'duration_secs': 0.390228} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.348693] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Reconfigured VM instance instance-0000000e to attach disk [datastore2] 6976b964-a8d3-4886-8aac-8d513e721018/6976b964-a8d3-4886-8aac-8d513e721018.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 663.351746] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b959846a-576c-463f-b5b0-5324368e297f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.362271] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for the task: (returnval){ [ 663.362271] env[62814]: value = "task-4293320" [ 663.362271] env[62814]: _type = "Task" [ 663.362271] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.375231] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293320, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.379880] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293319, 'name': Rename_Task, 'duration_secs': 0.26665} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.380224] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 663.380882] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-900f63e6-f71b-47c3-aacf-2c3cd75a4500 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.383041] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 663.383290] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b4fac1ae-21ab-4fd4-bdd5-f8d9d25d2023 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.389430] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 663.389430] env[62814]: value = "task-4293321" [ 663.389430] env[62814]: _type = "Task" [ 663.389430] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.393710] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 663.393710] env[62814]: value = "task-4293322" [ 663.393710] env[62814]: _type = "Task" [ 663.393710] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.400485] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293321, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.411647] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293322, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.501553] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796e8e52-c310-4203-871f-9e76f00bc52c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.524172] env[62814]: INFO nova.compute.manager [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Took 30.33 seconds to build instance. [ 663.525478] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40fc57c2-3267-4e59-9a6e-df6668031279 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.567122] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cc738a-87a4-43b7-a7a4-8da2ee5de81a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.576176] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65cfb53-dd56-4e19-900e-5ceaba3f5931 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.593112] env[62814]: DEBUG nova.compute.provider_tree [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 663.600984] env[62814]: DEBUG nova.network.neutron [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Updated VIF entry in instance network info cache for port 4615b448-a51a-4bfb-85fb-fd8f118e479d. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 663.602171] env[62814]: DEBUG nova.network.neutron [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Updating instance_info_cache with network_info: [{"id": "4615b448-a51a-4bfb-85fb-fd8f118e479d", "address": "fa:16:3e:97:b0:71", "network": {"id": "50951181-5761-45cf-9bd9-9f6766592c6b", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-505857443-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "df24d7d0c7c240ea960e83b4b809a6b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b911797-478d-4ee5-bce9-6f2f49014e94", "external-id": "nsx-vlan-transportzone-70", "segmentation_id": 70, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4615b448-a5", "ovs_interfaceid": "4615b448-a51a-4bfb-85fb-fd8f118e479d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.875281] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293320, 'name': Rename_Task, 'duration_secs': 0.221984} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.875799] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 663.875919] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7fa13fc5-3605-4857-a99a-6524d124768d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.883331] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for the task: (returnval){ [ 663.883331] env[62814]: value = "task-4293324" [ 663.883331] env[62814]: _type = "Task" [ 663.883331] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.892524] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293324, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.906824] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293321, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.912924] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293322, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.951298] env[62814]: DEBUG nova.network.neutron [req-d3ea7483-2455-4f0c-8879-8c85932ec2ef req-cf1643e4-668c-4bac-961b-8922b65e8b2f service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Updated VIF entry in instance network info cache for port 80990e09-7766-4c2a-937b-52c703caa252. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 663.951647] env[62814]: DEBUG nova.network.neutron [req-d3ea7483-2455-4f0c-8879-8c85932ec2ef req-cf1643e4-668c-4bac-961b-8922b65e8b2f service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Updating instance_info_cache with network_info: [{"id": "80990e09-7766-4c2a-937b-52c703caa252", "address": "fa:16:3e:92:a4:40", "network": {"id": "b4436b99-a4a3-46c9-918a-4e6f6f6bb55e", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-2020231604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d51711cd907404b8d0a48d904b599c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98d96b75-ac36-499a-adc2-130c8c1d55ca", "external-id": "nsx-vlan-transportzone-564", "segmentation_id": 564, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap80990e09-77", "ovs_interfaceid": "80990e09-7766-4c2a-937b-52c703caa252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.032246] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bcc55c04-4471-4ed4-9877-eda52e3245d5 tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.857s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 664.103755] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c7cce98-3436-4c4f-836f-8a214a9dfea8 req-18c8d844-5089-44a3-9174-0247700823ab service nova] Releasing lock "refresh_cache-6976b964-a8d3-4886-8aac-8d513e721018" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.132289] env[62814]: ERROR nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [req-a6683fd1-35b9-4717-b0e7-3f61e2520ecf] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-a6683fd1-35b9-4717-b0e7-3f61e2520ecf"}]} [ 664.151326] env[62814]: DEBUG nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 664.180475] env[62814]: DEBUG nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 664.180728] env[62814]: DEBUG nova.compute.provider_tree [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 664.200136] env[62814]: DEBUG nova.network.neutron [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Successfully updated port: a2e379d6-9cd3-4a22-85c2-72628174b4a6 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 664.202941] env[62814]: DEBUG nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 664.236573] env[62814]: DEBUG nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 664.409238] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293324, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.409680] env[62814]: DEBUG oslo_vmware.api [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293321, 'name': PowerOnVM_Task, 'duration_secs': 0.546537} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.410200] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 664.410724] env[62814]: INFO nova.compute.manager [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Took 7.02 seconds to spawn the instance on the hypervisor. [ 664.410724] env[62814]: DEBUG nova.compute.manager [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 664.411510] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299dea4b-15ce-4df2-bbe5-ae360c76a6f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.417911] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293322, 'name': CreateSnapshot_Task, 'duration_secs': 0.625006} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.421237] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 664.421873] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3155c3b5-cbe2-4809-a228-fe0de9c1d5b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.454899] env[62814]: DEBUG oslo_concurrency.lockutils [req-d3ea7483-2455-4f0c-8879-8c85932ec2ef req-cf1643e4-668c-4bac-961b-8922b65e8b2f service nova] Releasing lock "refresh_cache-5087b202-9bba-4489-823b-5d93cbf116e2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.538371] env[62814]: DEBUG nova.compute.manager [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 664.711946] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "refresh_cache-d9242042-6209-4b04-bf00-00dd04d9d6a0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.712506] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "refresh_cache-d9242042-6209-4b04-bf00-00dd04d9d6a0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 664.712506] env[62814]: DEBUG nova.network.neutron [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.746192] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8adeb0-99de-4e42-a5ba-1b06b745d110 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.758572] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e97a49b-a30f-4d7c-8669-409eeb851b95 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.810819] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c90cce-009b-4779-ad65-4ca4a77e8cbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.818569] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45275b51-25ac-4299-8e8d-4500703e379b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.835173] env[62814]: DEBUG nova.compute.provider_tree [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 664.902833] env[62814]: DEBUG oslo_vmware.api [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293324, 'name': PowerOnVM_Task, 'duration_secs': 0.577507} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.903175] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 664.903317] env[62814]: INFO nova.compute.manager [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Took 10.13 seconds to spawn the instance on the hypervisor. [ 664.904929] env[62814]: DEBUG nova.compute.manager [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 664.904929] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e8f64a-ddd5-4965-811a-97135e6b34b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.944362] env[62814]: INFO nova.compute.manager [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Took 26.06 seconds to build instance. [ 664.967082] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 664.967082] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a37acb37-4258-4ee7-a389-3332b2fb7d19 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.977679] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 664.977679] env[62814]: value = "task-4293325" [ 664.977679] env[62814]: _type = "Task" [ 664.977679] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.992565] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293325, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.019328] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.019580] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.075810] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.122696] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "64a475e4-6713-408b-a63a-a43b5fed5ec8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.122842] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "64a475e4-6713-408b-a63a-a43b5fed5ec8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.167308] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "961f313b-b43f-4531-8a4b-0a39421d6a34" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.167656] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "961f313b-b43f-4531-8a4b-0a39421d6a34" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.276366] env[62814]: DEBUG nova.network.neutron [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.367510] env[62814]: ERROR nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [req-7211ec90-861a-4004-89a7-e4053e0d778c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7211ec90-861a-4004-89a7-e4053e0d778c"}]} [ 665.393052] env[62814]: DEBUG nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 665.414568] env[62814]: DEBUG nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 665.414855] env[62814]: DEBUG nova.compute.provider_tree [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 665.436375] env[62814]: INFO nova.compute.manager [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Took 28.92 seconds to build instance. [ 665.445138] env[62814]: DEBUG nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: 4341bf54-74ba-4b7a-8fbf-155f2790d52b {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 665.454387] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4cb90-128d-4778-9551-1be0bfeebead tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lock "296f6c11-7108-42e6-8ada-5d8c08b00da6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.473s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.471619] env[62814]: DEBUG nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 665.494102] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293325, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.610252] env[62814]: DEBUG nova.network.neutron [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Updating instance_info_cache with network_info: [{"id": "a2e379d6-9cd3-4a22-85c2-72628174b4a6", "address": "fa:16:3e:3d:52:e0", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2e379d6-9c", "ovs_interfaceid": "a2e379d6-9cd3-4a22-85c2-72628174b4a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.739317] env[62814]: DEBUG nova.compute.manager [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Received event network-vif-plugged-a2e379d6-9cd3-4a22-85c2-72628174b4a6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 665.739317] env[62814]: DEBUG oslo_concurrency.lockutils [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] Acquiring lock "d9242042-6209-4b04-bf00-00dd04d9d6a0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.739317] env[62814]: DEBUG oslo_concurrency.lockutils [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] Lock "d9242042-6209-4b04-bf00-00dd04d9d6a0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.739317] env[62814]: DEBUG oslo_concurrency.lockutils [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] Lock "d9242042-6209-4b04-bf00-00dd04d9d6a0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.739981] env[62814]: DEBUG nova.compute.manager [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] No waiting events found dispatching network-vif-plugged-a2e379d6-9cd3-4a22-85c2-72628174b4a6 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 665.740450] env[62814]: WARNING nova.compute.manager [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Received unexpected event network-vif-plugged-a2e379d6-9cd3-4a22-85c2-72628174b4a6 for instance with vm_state building and task_state spawning. [ 665.740754] env[62814]: DEBUG nova.compute.manager [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Received event network-changed-a2e379d6-9cd3-4a22-85c2-72628174b4a6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 665.741088] env[62814]: DEBUG nova.compute.manager [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Refreshing instance network info cache due to event network-changed-a2e379d6-9cd3-4a22-85c2-72628174b4a6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 665.741741] env[62814]: DEBUG oslo_concurrency.lockutils [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] Acquiring lock "refresh_cache-d9242042-6209-4b04-bf00-00dd04d9d6a0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.945363] env[62814]: DEBUG oslo_concurrency.lockutils [None req-be551d7f-4e1e-4cb7-8a25-133e55d0a35f tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "6976b964-a8d3-4886-8aac-8d513e721018" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.430s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.959513] env[62814]: DEBUG nova.compute.manager [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 666.002367] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293325, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.115646] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "refresh_cache-d9242042-6209-4b04-bf00-00dd04d9d6a0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.116058] env[62814]: DEBUG nova.compute.manager [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Instance network_info: |[{"id": "a2e379d6-9cd3-4a22-85c2-72628174b4a6", "address": "fa:16:3e:3d:52:e0", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2e379d6-9c", "ovs_interfaceid": "a2e379d6-9cd3-4a22-85c2-72628174b4a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 666.116246] env[62814]: DEBUG oslo_concurrency.lockutils [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] Acquired lock "refresh_cache-d9242042-6209-4b04-bf00-00dd04d9d6a0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 666.116412] env[62814]: DEBUG nova.network.neutron [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Refreshing network info cache for port a2e379d6-9cd3-4a22-85c2-72628174b4a6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 666.119157] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:52:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c621a9c-66f5-426a-8aab-bd8b2e912106', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a2e379d6-9cd3-4a22-85c2-72628174b4a6', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 666.128540] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 666.129282] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 666.130236] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2f141c6b-159d-4cb5-aee9-e2dc51e8fd7c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.148069] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43d4380-047b-4e1d-ac0c-9d37bca37220 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.158470] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba446fc-090e-495b-9f54-64fbb92f2cf9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.161469] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 666.161469] env[62814]: value = "task-4293327" [ 666.161469] env[62814]: _type = "Task" [ 666.161469] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.193257] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4234c1ba-3761-4625-a474-1961c7dd5963 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.199639] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293327, 'name': CreateVM_Task} progress is 15%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.204701] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f75f4cc-7214-41f2-b50a-99e2fccb07a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.220305] env[62814]: DEBUG nova.compute.provider_tree [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 666.448690] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 666.495099] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293325, 'name': CloneVM_Task, 'duration_secs': 1.421801} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.496047] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Created linked-clone VM from snapshot [ 666.499422] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a7a0d6-9ce8-4363-9af6-9008f209af47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.508789] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Uploading image e7328b9e-7c4e-41a4-ade1-d439df2073ed {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 666.512225] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.561364] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 666.561364] env[62814]: value = "vm-845598" [ 666.561364] env[62814]: _type = "VirtualMachine" [ 666.561364] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 666.561704] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-ea9e1f7b-db9e-4807-9d78-a2e6379cd066 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.568671] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lease: (returnval){ [ 666.568671] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5230286c-7c01-bb5e-f0af-3fb36fac7892" [ 666.568671] env[62814]: _type = "HttpNfcLease" [ 666.568671] env[62814]: } obtained for exporting VM: (result){ [ 666.568671] env[62814]: value = "vm-845598" [ 666.568671] env[62814]: _type = "VirtualMachine" [ 666.568671] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 666.568943] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the lease: (returnval){ [ 666.568943] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5230286c-7c01-bb5e-f0af-3fb36fac7892" [ 666.568943] env[62814]: _type = "HttpNfcLease" [ 666.568943] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 666.579851] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 666.579851] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5230286c-7c01-bb5e-f0af-3fb36fac7892" [ 666.579851] env[62814]: _type = "HttpNfcLease" [ 666.579851] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 666.676619] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293327, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.805952] env[62814]: DEBUG nova.scheduler.client.report [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 37 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 666.806256] env[62814]: DEBUG nova.compute.provider_tree [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 37 to 38 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 666.806436] env[62814]: DEBUG nova.compute.provider_tree [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 666.973316] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 667.083794] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 667.083794] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5230286c-7c01-bb5e-f0af-3fb36fac7892" [ 667.083794] env[62814]: _type = "HttpNfcLease" [ 667.083794] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 667.084762] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 667.084762] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5230286c-7c01-bb5e-f0af-3fb36fac7892" [ 667.084762] env[62814]: _type = "HttpNfcLease" [ 667.084762] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 667.085340] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1a2647-56e7-4fa2-9d0b-28ee6d975407 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.096519] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cf621-8160-4031-40f2-0c893e2ec830/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 667.096719] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cf621-8160-4031-40f2-0c893e2ec830/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 667.178366] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293327, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.212895] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3dc60038-7b9c-4889-bf0b-eb557f4579eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.244483] env[62814]: DEBUG nova.compute.manager [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 667.245474] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b9d0c9e-cad9-47bb-a7c9-512a64dc454c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.257217] env[62814]: DEBUG nova.network.neutron [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Updated VIF entry in instance network info cache for port a2e379d6-9cd3-4a22-85c2-72628174b4a6. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 667.257558] env[62814]: DEBUG nova.network.neutron [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Updating instance_info_cache with network_info: [{"id": "a2e379d6-9cd3-4a22-85c2-72628174b4a6", "address": "fa:16:3e:3d:52:e0", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2e379d6-9c", "ovs_interfaceid": "a2e379d6-9cd3-4a22-85c2-72628174b4a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.312187] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 5.236s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.314413] env[62814]: DEBUG nova.compute.manager [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 667.315950] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.036s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 667.317451] env[62814]: INFO nova.compute.claims [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.675488] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293327, 'name': CreateVM_Task, 'duration_secs': 1.414894} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.675818] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 667.676571] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.676950] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 667.677649] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 667.678761] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70913d67-2ae1-4878-a0c8-cbb3b764c5b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.686863] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 667.686863] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52399cb3-3eaa-0479-f3c2-78329d333626" [ 667.686863] env[62814]: _type = "Task" [ 667.686863] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.698291] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52399cb3-3eaa-0479-f3c2-78329d333626, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.762535] env[62814]: DEBUG oslo_concurrency.lockutils [req-0f1cb3dd-969e-4d61-8f24-4939cfb7ad1a req-c385d794-1be0-434b-b0c9-0383aae9d7be service nova] Releasing lock "refresh_cache-d9242042-6209-4b04-bf00-00dd04d9d6a0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 667.768459] env[62814]: INFO nova.compute.manager [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] instance snapshotting [ 667.769055] env[62814]: DEBUG nova.objects.instance [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lazy-loading 'flavor' on Instance uuid 296f6c11-7108-42e6-8ada-5d8c08b00da6 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 667.823933] env[62814]: DEBUG nova.compute.utils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 667.829663] env[62814]: DEBUG nova.compute.manager [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 667.829790] env[62814]: DEBUG nova.network.neutron [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 667.995286] env[62814]: DEBUG nova.policy [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3cbeb7b7e7fc46ffac95e891184f56b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97d11f49b6ad4c82956efa254d8995ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 668.022675] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquiring lock "004fd137-4902-4313-a6f7-6c83cd76743d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 668.022982] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lock "004fd137-4902-4313-a6f7-6c83cd76743d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.198611] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52399cb3-3eaa-0479-f3c2-78329d333626, 'name': SearchDatastore_Task, 'duration_secs': 0.012096} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.198932] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.199217] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 668.199459] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.199945] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.199945] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 668.200190] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-741d7053-bb3c-41b2-8dcd-c4735f539d4b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.209992] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 668.210231] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 668.211028] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a06f09de-21df-4ff1-a5a1-016ebfad76f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.225038] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 668.225038] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527f7df2-766b-ddd7-75be-39841d94a40b" [ 668.225038] env[62814]: _type = "Task" [ 668.225038] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.234324] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527f7df2-766b-ddd7-75be-39841d94a40b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.277055] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb5bf89-90d2-4c84-945c-53d3aa01cba1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.306117] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b92e5874-f6fc-44f5-9646-d69b65141c0d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.331038] env[62814]: DEBUG nova.compute.manager [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 668.746678] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527f7df2-766b-ddd7-75be-39841d94a40b, 'name': SearchDatastore_Task, 'duration_secs': 0.01012} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.748412] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83b91c70-8ee7-40ad-aedf-d5a81ae5279b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.755069] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 668.755069] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52751591-3277-1f57-a407-fccf04c7e4d6" [ 668.755069] env[62814]: _type = "Task" [ 668.755069] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.767120] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52751591-3277-1f57-a407-fccf04c7e4d6, 'name': SearchDatastore_Task, 'duration_secs': 0.009767} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.767369] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.767617] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d9242042-6209-4b04-bf00-00dd04d9d6a0/d9242042-6209-4b04-bf00-00dd04d9d6a0.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 668.767859] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b0e2af90-e5f0-4af2-be41-9a7323cef677 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.777750] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 668.777750] env[62814]: value = "task-4293329" [ 668.777750] env[62814]: _type = "Task" [ 668.777750] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.784364] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293329, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.819565] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 668.820289] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-10e4e3d0-f34b-41a4-9551-104eb659579c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.838876] env[62814]: DEBUG oslo_vmware.api [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 668.838876] env[62814]: value = "task-4293330" [ 668.838876] env[62814]: _type = "Task" [ 668.838876] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.849485] env[62814]: DEBUG oslo_vmware.api [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293330, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.905239] env[62814]: DEBUG nova.network.neutron [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Successfully created port: 05dc355c-bbf9-44a2-80dc-4a730699b8f8 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.934277] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a028c2e6-f143-47c6-8b28-bc823e1241ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.944912] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0303cdb9-a131-40df-8c2c-081be41400a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.991467] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7da2a07f-11ac-41ab-9ef9-d8ef8f5d5bf8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.000688] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac683dc5-7e96-4845-9146-7317f7a0d8d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.018095] env[62814]: DEBUG nova.compute.provider_tree [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 669.033368] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquiring lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.033638] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.293012] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293329, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.344885] env[62814]: DEBUG nova.compute.manager [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 669.351321] env[62814]: DEBUG oslo_vmware.api [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293330, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.384907] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 669.385541] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.385541] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 669.385867] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.386049] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 669.386273] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 669.386560] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 669.386740] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 669.386917] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 669.387105] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 669.387288] env[62814]: DEBUG nova.virt.hardware [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 669.389018] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78232335-7380-4b8f-8746-d563e12a64e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.396501] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c899811-a932-437d-8c50-9c47e213bfe3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.499549] env[62814]: DEBUG nova.compute.manager [req-14931c61-a781-4a69-bb1d-a71c1d93a7f3 req-13c88997-137e-4dbd-b934-b3b84cd73901 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Received event network-changed-25fab244-2cae-416b-86e1-f25d1ea5aae7 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 669.499753] env[62814]: DEBUG nova.compute.manager [req-14931c61-a781-4a69-bb1d-a71c1d93a7f3 req-13c88997-137e-4dbd-b934-b3b84cd73901 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Refreshing instance network info cache due to event network-changed-25fab244-2cae-416b-86e1-f25d1ea5aae7. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 669.499958] env[62814]: DEBUG oslo_concurrency.lockutils [req-14931c61-a781-4a69-bb1d-a71c1d93a7f3 req-13c88997-137e-4dbd-b934-b3b84cd73901 service nova] Acquiring lock "refresh_cache-3d3f32fc-276a-49be-b471-01a5d6fc5069" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.500116] env[62814]: DEBUG oslo_concurrency.lockutils [req-14931c61-a781-4a69-bb1d-a71c1d93a7f3 req-13c88997-137e-4dbd-b934-b3b84cd73901 service nova] Acquired lock "refresh_cache-3d3f32fc-276a-49be-b471-01a5d6fc5069" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 669.500272] env[62814]: DEBUG nova.network.neutron [req-14931c61-a781-4a69-bb1d-a71c1d93a7f3 req-13c88997-137e-4dbd-b934-b3b84cd73901 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Refreshing network info cache for port 25fab244-2cae-416b-86e1-f25d1ea5aae7 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 669.553277] env[62814]: ERROR nova.scheduler.client.report [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [req-2f935fdf-3e16-4d7d-8c9b-8c74211119c3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2f935fdf-3e16-4d7d-8c9b-8c74211119c3"}]} [ 669.572889] env[62814]: DEBUG nova.scheduler.client.report [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 669.595297] env[62814]: DEBUG nova.scheduler.client.report [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 669.595297] env[62814]: DEBUG nova.compute.provider_tree [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 669.621899] env[62814]: DEBUG nova.scheduler.client.report [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 669.644470] env[62814]: DEBUG nova.scheduler.client.report [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 669.737457] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquiring lock "296f6c11-7108-42e6-8ada-5d8c08b00da6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.737712] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lock "296f6c11-7108-42e6-8ada-5d8c08b00da6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.737918] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquiring lock "296f6c11-7108-42e6-8ada-5d8c08b00da6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 669.740266] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lock "296f6c11-7108-42e6-8ada-5d8c08b00da6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 669.741059] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lock "296f6c11-7108-42e6-8ada-5d8c08b00da6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.002s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 669.744149] env[62814]: INFO nova.compute.manager [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Terminating instance [ 669.793069] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293329, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546149} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.793335] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d9242042-6209-4b04-bf00-00dd04d9d6a0/d9242042-6209-4b04-bf00-00dd04d9d6a0.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 669.793462] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 669.793980] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e8668a3-a7f9-4f9e-a812-f81bec8cd374 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.801364] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 669.801364] env[62814]: value = "task-4293331" [ 669.801364] env[62814]: _type = "Task" [ 669.801364] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.810464] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293331, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.853948] env[62814]: DEBUG oslo_vmware.api [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293330, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.178946] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5199f064-4a5b-4858-8122-32598d14a44d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.187031] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26c7de2-f22e-43bc-b486-ef22c5793619 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.218853] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2acf8e19-14c2-4a3e-993b-cf7add35a4b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.227076] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780bd495-e0a9-4845-8a9d-dfe3c1d17b40 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.242596] env[62814]: DEBUG nova.compute.provider_tree [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.248706] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquiring lock "refresh_cache-296f6c11-7108-42e6-8ada-5d8c08b00da6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.248880] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquired lock "refresh_cache-296f6c11-7108-42e6-8ada-5d8c08b00da6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 670.249065] env[62814]: DEBUG nova.network.neutron [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 670.314811] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293331, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075912} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.315121] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 670.315975] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f687ecec-4296-4f79-b74b-536ddbe59482 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.345505] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] d9242042-6209-4b04-bf00-00dd04d9d6a0/d9242042-6209-4b04-bf00-00dd04d9d6a0.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 670.345505] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6526ac4d-d714-4601-844e-d49fd6efd4ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.372271] env[62814]: DEBUG oslo_vmware.api [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293330, 'name': CreateSnapshot_Task, 'duration_secs': 1.039993} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.373862] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 670.374189] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 670.374189] env[62814]: value = "task-4293332" [ 670.374189] env[62814]: _type = "Task" [ 670.374189] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.374808] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a6103e-1c35-4327-9da1-40027536b9f3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.387743] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293332, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.514182] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18ce047f-2199-401c-8155-37e42fd033e1 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquiring lock "interface-6976b964-a8d3-4886-8aac-8d513e721018-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 670.514480] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18ce047f-2199-401c-8155-37e42fd033e1 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "interface-6976b964-a8d3-4886-8aac-8d513e721018-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 670.514937] env[62814]: DEBUG nova.objects.instance [None req-18ce047f-2199-401c-8155-37e42fd033e1 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lazy-loading 'flavor' on Instance uuid 6976b964-a8d3-4886-8aac-8d513e721018 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 670.745449] env[62814]: DEBUG nova.scheduler.client.report [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 670.781789] env[62814]: DEBUG nova.network.neutron [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.888906] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293332, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.896670] env[62814]: DEBUG nova.compute.manager [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Instance disappeared during snapshot {{(pid=62814) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4595}} [ 671.023085] env[62814]: DEBUG nova.objects.instance [None req-18ce047f-2199-401c-8155-37e42fd033e1 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lazy-loading 'pci_requests' on Instance uuid 6976b964-a8d3-4886-8aac-8d513e721018 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 671.070028] env[62814]: DEBUG nova.compute.manager [None req-c44e50e4-49a7-40e8-8d30-98da3e4a1acf tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Found 0 images (rotation: 2) {{(pid=62814) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5017}} [ 671.095720] env[62814]: DEBUG nova.network.neutron [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.222592] env[62814]: DEBUG nova.network.neutron [req-14931c61-a781-4a69-bb1d-a71c1d93a7f3 req-13c88997-137e-4dbd-b934-b3b84cd73901 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Updated VIF entry in instance network info cache for port 25fab244-2cae-416b-86e1-f25d1ea5aae7. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 671.222972] env[62814]: DEBUG nova.network.neutron [req-14931c61-a781-4a69-bb1d-a71c1d93a7f3 req-13c88997-137e-4dbd-b934-b3b84cd73901 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Updating instance_info_cache with network_info: [{"id": "25fab244-2cae-416b-86e1-f25d1ea5aae7", "address": "fa:16:3e:b8:89:0d", "network": {"id": "8a6a9999-88fc-4c83-b519-b2b9a657ab49", "bridge": "br-int", "label": "tempest-VolumesAssistedSnapshotsTest-1596604004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.218", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "07ee864ce77f44de84624ee579fb5316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78340140-126f-4ef8-a340-debaa64da3e5", "external-id": "nsx-vlan-transportzone-648", "segmentation_id": 648, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25fab244-2c", "ovs_interfaceid": "25fab244-2cae-416b-86e1-f25d1ea5aae7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.254251] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.935s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 671.254251] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 671.257439] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.894s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 671.261449] env[62814]: INFO nova.compute.claims [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.391288] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293332, 'name': ReconfigVM_Task, 'duration_secs': 0.749231} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.391440] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Reconfigured VM instance instance-00000010 to attach disk [datastore2] d9242042-6209-4b04-bf00-00dd04d9d6a0/d9242042-6209-4b04-bf00-00dd04d9d6a0.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 671.392120] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae028835-b41b-4e4d-8e02-817b3de0a869 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.401392] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 671.401392] env[62814]: value = "task-4293333" [ 671.401392] env[62814]: _type = "Task" [ 671.401392] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.411236] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293333, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.532059] env[62814]: DEBUG nova.objects.base [None req-18ce047f-2199-401c-8155-37e42fd033e1 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Object Instance<6976b964-a8d3-4886-8aac-8d513e721018> lazy-loaded attributes: flavor,pci_requests {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 671.533949] env[62814]: DEBUG nova.network.neutron [None req-18ce047f-2199-401c-8155-37e42fd033e1 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 671.604109] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Releasing lock "refresh_cache-296f6c11-7108-42e6-8ada-5d8c08b00da6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 671.604608] env[62814]: DEBUG nova.compute.manager [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 671.604848] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 671.605748] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67fcad94-30c4-41b0-a5e6-bb89ce6a2735 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.618334] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 671.618824] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-192dd01c-bf12-4f00-910f-06db6e9c119e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.630902] env[62814]: DEBUG oslo_vmware.api [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 671.630902] env[62814]: value = "task-4293334" [ 671.630902] env[62814]: _type = "Task" [ 671.630902] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.641726] env[62814]: DEBUG oslo_vmware.api [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293334, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.728890] env[62814]: DEBUG oslo_concurrency.lockutils [req-14931c61-a781-4a69-bb1d-a71c1d93a7f3 req-13c88997-137e-4dbd-b934-b3b84cd73901 service nova] Releasing lock "refresh_cache-3d3f32fc-276a-49be-b471-01a5d6fc5069" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 671.768565] env[62814]: DEBUG nova.compute.utils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 671.776677] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 671.776897] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 671.854848] env[62814]: DEBUG nova.policy [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24c9d7684648493c8e4beed961a51d3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eadee196b6264ba987043438ccdce5f8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 671.914338] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293333, 'name': Rename_Task, 'duration_secs': 0.290701} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.915335] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 671.915750] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6af80648-b819-4774-b7f5-0851408cec3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.926774] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 671.926774] env[62814]: value = "task-4293335" [ 671.926774] env[62814]: _type = "Task" [ 671.926774] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.939825] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293335, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.950636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-18ce047f-2199-401c-8155-37e42fd033e1 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "interface-6976b964-a8d3-4886-8aac-8d513e721018-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.436s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 671.965121] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Acquiring lock "bd0933ca-aab7-4dd4-a570-1a58a720f377" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 671.965121] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lock "bd0933ca-aab7-4dd4-a570-1a58a720f377" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 672.143970] env[62814]: DEBUG oslo_vmware.api [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293334, 'name': PowerOffVM_Task, 'duration_secs': 0.230583} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.144160] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 672.144411] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 672.144684] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-262feba6-cae7-4586-a84e-5aa263a649a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.187630] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 672.188270] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 672.188525] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Deleting the datastore file [datastore2] 296f6c11-7108-42e6-8ada-5d8c08b00da6 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 672.188862] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f8933bd4-8da1-42a6-8347-232768bcd64c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.200061] env[62814]: DEBUG oslo_vmware.api [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for the task: (returnval){ [ 672.200061] env[62814]: value = "task-4293337" [ 672.200061] env[62814]: _type = "Task" [ 672.200061] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.210919] env[62814]: DEBUG oslo_vmware.api [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293337, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.277187] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 672.330944] env[62814]: DEBUG nova.network.neutron [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Successfully updated port: 05dc355c-bbf9-44a2-80dc-4a730699b8f8 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 672.399673] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Successfully created port: fb397a6f-5617-48a5-b5f6-30d9be90c5d4 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 672.446413] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293335, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.714810] env[62814]: DEBUG oslo_vmware.api [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Task: {'id': task-4293337, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211238} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.718354] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 672.718354] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 672.718673] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 672.718939] env[62814]: INFO nova.compute.manager [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Took 1.11 seconds to destroy the instance on the hypervisor. [ 672.719829] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 672.720505] env[62814]: DEBUG nova.compute.manager [-] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 672.720505] env[62814]: DEBUG nova.network.neutron [-] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.835449] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "refresh_cache-98d134b4-b4ca-4247-a638-ad5c24a694e5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.837355] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquired lock "refresh_cache-98d134b4-b4ca-4247-a638-ad5c24a694e5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.837355] env[62814]: DEBUG nova.network.neutron [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 672.864404] env[62814]: DEBUG nova.network.neutron [-] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.946888] env[62814]: DEBUG oslo_vmware.api [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293335, 'name': PowerOnVM_Task, 'duration_secs': 0.734989} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.946888] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 672.947258] env[62814]: INFO nova.compute.manager [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Took 10.95 seconds to spawn the instance on the hypervisor. [ 672.947258] env[62814]: DEBUG nova.compute.manager [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 672.948663] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e9ad23-3ca3-421a-9b50-81dc058234b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.996832] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da3508c-36b3-4dd7-94f2-d6769b937474 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.006557] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151e982f-d5b5-4b0c-8ad1-00f245ef42ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.047307] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8e623bb-1a1b-4289-bdb9-f7a8e04b63d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.061023] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221ed09f-93b0-48a2-b67f-c06b38909ee6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.084066] env[62814]: DEBUG nova.compute.provider_tree [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.290708] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 673.320639] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 673.321113] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 673.321113] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 673.321260] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 673.321370] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 673.321542] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 673.323541] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 673.323541] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 673.323541] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 673.323856] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 673.324106] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 673.324882] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ae97f9-e56a-48e7-8edc-12d78544f3d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.345177] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e48849d-9145-4c94-92c5-e78ab1923c3f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.372028] env[62814]: DEBUG nova.network.neutron [-] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.434438] env[62814]: DEBUG nova.network.neutron [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.473264] env[62814]: DEBUG nova.compute.manager [req-4e12ee95-cc2b-4e24-9ccc-713cad3e0192 req-5ce42206-a26c-4954-99d1-db40661e25ac service nova] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Received event network-vif-plugged-05dc355c-bbf9-44a2-80dc-4a730699b8f8 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 673.474157] env[62814]: DEBUG oslo_concurrency.lockutils [req-4e12ee95-cc2b-4e24-9ccc-713cad3e0192 req-5ce42206-a26c-4954-99d1-db40661e25ac service nova] Acquiring lock "98d134b4-b4ca-4247-a638-ad5c24a694e5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 673.474833] env[62814]: DEBUG oslo_concurrency.lockutils [req-4e12ee95-cc2b-4e24-9ccc-713cad3e0192 req-5ce42206-a26c-4954-99d1-db40661e25ac service nova] Lock "98d134b4-b4ca-4247-a638-ad5c24a694e5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 673.475017] env[62814]: DEBUG oslo_concurrency.lockutils [req-4e12ee95-cc2b-4e24-9ccc-713cad3e0192 req-5ce42206-a26c-4954-99d1-db40661e25ac service nova] Lock "98d134b4-b4ca-4247-a638-ad5c24a694e5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 673.475237] env[62814]: DEBUG nova.compute.manager [req-4e12ee95-cc2b-4e24-9ccc-713cad3e0192 req-5ce42206-a26c-4954-99d1-db40661e25ac service nova] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] No waiting events found dispatching network-vif-plugged-05dc355c-bbf9-44a2-80dc-4a730699b8f8 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 673.475407] env[62814]: WARNING nova.compute.manager [req-4e12ee95-cc2b-4e24-9ccc-713cad3e0192 req-5ce42206-a26c-4954-99d1-db40661e25ac service nova] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Received unexpected event network-vif-plugged-05dc355c-bbf9-44a2-80dc-4a730699b8f8 for instance with vm_state building and task_state spawning. [ 673.477827] env[62814]: INFO nova.compute.manager [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Took 31.67 seconds to build instance. [ 673.587031] env[62814]: DEBUG nova.scheduler.client.report [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 673.631022] env[62814]: DEBUG nova.network.neutron [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Updating instance_info_cache with network_info: [{"id": "05dc355c-bbf9-44a2-80dc-4a730699b8f8", "address": "fa:16:3e:7c:07:98", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05dc355c-bb", "ovs_interfaceid": "05dc355c-bbf9-44a2-80dc-4a730699b8f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.880141] env[62814]: INFO nova.compute.manager [-] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Took 1.16 seconds to deallocate network for instance. [ 673.986476] env[62814]: DEBUG oslo_concurrency.lockutils [None req-789e90b7-bfac-4053-9f2f-065219dfa7c8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "d9242042-6209-4b04-bf00-00dd04d9d6a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.684s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.094734] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.837s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.095323] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 674.099241] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 25.026s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.135350] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Releasing lock "refresh_cache-98d134b4-b4ca-4247-a638-ad5c24a694e5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 674.135350] env[62814]: DEBUG nova.compute.manager [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Instance network_info: |[{"id": "05dc355c-bbf9-44a2-80dc-4a730699b8f8", "address": "fa:16:3e:7c:07:98", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05dc355c-bb", "ovs_interfaceid": "05dc355c-bbf9-44a2-80dc-4a730699b8f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 674.135739] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7c:07:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '05dc355c-bbf9-44a2-80dc-4a730699b8f8', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 674.151264] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 674.152033] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 674.159825] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4c959931-a3ce-42f7-96bd-f1ac4aec42a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.192762] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 674.192762] env[62814]: value = "task-4293338" [ 674.192762] env[62814]: _type = "Task" [ 674.192762] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.202828] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293338, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.392049] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.428884] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Successfully updated port: fb397a6f-5617-48a5-b5f6-30d9be90c5d4 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 674.486793] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 674.521926] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquiring lock "6976b964-a8d3-4886-8aac-8d513e721018" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.522310] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "6976b964-a8d3-4886-8aac-8d513e721018" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.522543] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquiring lock "6976b964-a8d3-4886-8aac-8d513e721018-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 674.522733] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "6976b964-a8d3-4886-8aac-8d513e721018-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 674.522901] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "6976b964-a8d3-4886-8aac-8d513e721018-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.525557] env[62814]: INFO nova.compute.manager [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Terminating instance [ 674.603678] env[62814]: DEBUG nova.compute.utils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 674.612861] env[62814]: INFO nova.compute.claims [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 674.616766] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 674.616766] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 674.676998] env[62814]: DEBUG nova.policy [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24c9d7684648493c8e4beed961a51d3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eadee196b6264ba987043438ccdce5f8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 674.705386] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293338, 'name': CreateVM_Task, 'duration_secs': 0.400087} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.705505] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 674.706123] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.706278] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.706709] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 674.706850] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e733a79b-5bb2-4563-9c17-bf658520775d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.714307] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 674.714307] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fadeac-6bc1-3907-1197-89d0797c37c1" [ 674.714307] env[62814]: _type = "Task" [ 674.714307] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.732265] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fadeac-6bc1-3907-1197-89d0797c37c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.931731] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "refresh_cache-a36ad785-2f33-4dbc-bc82-ab4a35020b0f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.931869] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "refresh_cache-a36ad785-2f33-4dbc-bc82-ab4a35020b0f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.933252] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.034448] env[62814]: DEBUG nova.compute.manager [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 675.034448] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 675.034448] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e88b9a4a-8614-4104-a6e5-0da9e9a928f3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.038695] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 675.047159] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 675.047790] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad9d17c4-5e08-4ef7-b9b2-06bcaa7ce212 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.058754] env[62814]: DEBUG oslo_vmware.api [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for the task: (returnval){ [ 675.058754] env[62814]: value = "task-4293339" [ 675.058754] env[62814]: _type = "Task" [ 675.058754] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.071105] env[62814]: DEBUG oslo_vmware.api [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293339, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.121225] env[62814]: INFO nova.compute.resource_tracker [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating resource usage from migration 126a427f-3696-4882-ac2f-5c3e4dbfe77f [ 675.127688] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 675.236805] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fadeac-6bc1-3907-1197-89d0797c37c1, 'name': SearchDatastore_Task, 'duration_secs': 0.013922} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.237161] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 675.237406] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 675.238020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.238020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 675.238020] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 675.238310] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76e3d1ce-c2e5-450c-b5d1-2afbd2de8fe0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.250368] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 675.250615] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 675.254678] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e336f93c-1ba2-4e7a-bd88-a6ea8557026a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.262053] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 675.262053] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527a4902-b013-7928-9e10-6f49e9426fd8" [ 675.262053] env[62814]: _type = "Task" [ 675.262053] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.274850] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527a4902-b013-7928-9e10-6f49e9426fd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.299154] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Successfully created port: 04ddd249-262e-42ae-95d7-a32509b50e6d {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 675.513334] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquiring lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 675.513950] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.518980] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquiring lock "369cd937-4c18-4068-ae59-70a1d585094b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 675.518980] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Lock "369cd937-4c18-4068-ae59-70a1d585094b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.572434] env[62814]: DEBUG oslo_vmware.api [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293339, 'name': PowerOffVM_Task, 'duration_secs': 0.273952} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.572743] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.574784] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 675.575494] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 675.575494] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-94605631-952c-409e-a8e8-4546776d4d3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.652236] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 675.652543] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 675.652657] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Deleting the datastore file [datastore2] 6976b964-a8d3-4886-8aac-8d513e721018 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 675.653793] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8cb2bf20-88ab-4a81-a349-149d06cb8b0e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.667031] env[62814]: DEBUG oslo_vmware.api [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for the task: (returnval){ [ 675.667031] env[62814]: value = "task-4293341" [ 675.667031] env[62814]: _type = "Task" [ 675.667031] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.677736] env[62814]: DEBUG oslo_vmware.api [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293341, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.679879] env[62814]: DEBUG nova.compute.manager [req-a6741a62-0956-4eb3-98d6-86d7ac915e7b req-11cd5449-16e3-4fd1-abe2-a77102885106 service nova] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Received event network-vif-plugged-fb397a6f-5617-48a5-b5f6-30d9be90c5d4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 675.680146] env[62814]: DEBUG oslo_concurrency.lockutils [req-a6741a62-0956-4eb3-98d6-86d7ac915e7b req-11cd5449-16e3-4fd1-abe2-a77102885106 service nova] Acquiring lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 675.680478] env[62814]: DEBUG oslo_concurrency.lockutils [req-a6741a62-0956-4eb3-98d6-86d7ac915e7b req-11cd5449-16e3-4fd1-abe2-a77102885106 service nova] Lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.680731] env[62814]: DEBUG oslo_concurrency.lockutils [req-a6741a62-0956-4eb3-98d6-86d7ac915e7b req-11cd5449-16e3-4fd1-abe2-a77102885106 service nova] Lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.681041] env[62814]: DEBUG nova.compute.manager [req-a6741a62-0956-4eb3-98d6-86d7ac915e7b req-11cd5449-16e3-4fd1-abe2-a77102885106 service nova] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] No waiting events found dispatching network-vif-plugged-fb397a6f-5617-48a5-b5f6-30d9be90c5d4 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 675.681427] env[62814]: WARNING nova.compute.manager [req-a6741a62-0956-4eb3-98d6-86d7ac915e7b req-11cd5449-16e3-4fd1-abe2-a77102885106 service nova] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Received unexpected event network-vif-plugged-fb397a6f-5617-48a5-b5f6-30d9be90c5d4 for instance with vm_state building and task_state spawning. [ 675.774533] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527a4902-b013-7928-9e10-6f49e9426fd8, 'name': SearchDatastore_Task, 'duration_secs': 0.015946} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.775719] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c52646f-df27-4d80-8f55-e32f62391d8b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.783233] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 675.783233] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ea789b-7f53-1b92-6eea-68f5f7c19c38" [ 675.783233] env[62814]: _type = "Task" [ 675.783233] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.795555] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ea789b-7f53-1b92-6eea-68f5f7c19c38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.886240] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f51f6e44-b819-4e7e-ac0c-41da5bfddc7c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.894973] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff00b7b-bb72-48da-bf13-865eff12d0e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.941240] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83232e32-67dc-4d84-82ea-8ca846e9bf64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.955297] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7830b04-2640-4337-90bf-5dadd90c5e08 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.962264] env[62814]: INFO nova.compute.manager [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Rescuing [ 675.962686] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "refresh_cache-d9242042-6209-4b04-bf00-00dd04d9d6a0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.962686] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "refresh_cache-d9242042-6209-4b04-bf00-00dd04d9d6a0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 675.962829] env[62814]: DEBUG nova.network.neutron [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.976272] env[62814]: DEBUG nova.compute.provider_tree [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 676.116482] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Updating instance_info_cache with network_info: [{"id": "fb397a6f-5617-48a5-b5f6-30d9be90c5d4", "address": "fa:16:3e:08:4a:69", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb397a6f-56", "ovs_interfaceid": "fb397a6f-5617-48a5-b5f6-30d9be90c5d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.140812] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 676.171646] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 676.171646] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 676.171646] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 676.172077] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 676.172077] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 676.172077] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 676.172077] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 676.172316] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 676.172543] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 676.172716] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 676.173153] env[62814]: DEBUG nova.virt.hardware [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 676.175327] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1bc443-03c8-4b23-9582-ff164301c07a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.190624] env[62814]: DEBUG oslo_vmware.api [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Task: {'id': task-4293341, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.207734} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.190977] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 676.191277] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 676.191624] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 676.191731] env[62814]: INFO nova.compute.manager [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Took 1.16 seconds to destroy the instance on the hypervisor. [ 676.191937] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 676.195042] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07366167-8c87-40da-9981-044baa01159b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.198381] env[62814]: DEBUG nova.compute.manager [-] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 676.198487] env[62814]: DEBUG nova.network.neutron [-] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.298446] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ea789b-7f53-1b92-6eea-68f5f7c19c38, 'name': SearchDatastore_Task, 'duration_secs': 0.015151} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.298771] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 676.299047] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98d134b4-b4ca-4247-a638-ad5c24a694e5/98d134b4-b4ca-4247-a638-ad5c24a694e5.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 676.299315] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cff9fea3-3f7a-4029-a3d6-9108050019d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.307529] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 676.307529] env[62814]: value = "task-4293342" [ 676.307529] env[62814]: _type = "Task" [ 676.307529] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.318232] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293342, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.510380] env[62814]: ERROR nova.scheduler.client.report [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [req-f8bce28b-2881-4d77-9d89-288842e426c2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f8bce28b-2881-4d77-9d89-288842e426c2"}]} [ 676.533659] env[62814]: DEBUG nova.scheduler.client.report [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 676.550132] env[62814]: DEBUG nova.scheduler.client.report [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 676.550390] env[62814]: DEBUG nova.compute.provider_tree [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 676.566268] env[62814]: DEBUG nova.scheduler.client.report [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 676.592568] env[62814]: DEBUG nova.scheduler.client.report [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 676.624239] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "refresh_cache-a36ad785-2f33-4dbc-bc82-ab4a35020b0f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 676.624983] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Instance network_info: |[{"id": "fb397a6f-5617-48a5-b5f6-30d9be90c5d4", "address": "fa:16:3e:08:4a:69", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb397a6f-56", "ovs_interfaceid": "fb397a6f-5617-48a5-b5f6-30d9be90c5d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 676.629831] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:08:4a:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fb397a6f-5617-48a5-b5f6-30d9be90c5d4', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 676.646719] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Creating folder: Project (eadee196b6264ba987043438ccdce5f8). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 676.651791] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-05009a98-e78c-4860-a334-0a4bab81993a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.661377] env[62814]: DEBUG nova.compute.manager [req-e1db582f-f8d2-4cf0-a33c-46704589f271 req-30ca30be-06dc-4810-95c9-73fb075eb696 service nova] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Received event network-changed-05dc355c-bbf9-44a2-80dc-4a730699b8f8 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 676.661377] env[62814]: DEBUG nova.compute.manager [req-e1db582f-f8d2-4cf0-a33c-46704589f271 req-30ca30be-06dc-4810-95c9-73fb075eb696 service nova] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Refreshing instance network info cache due to event network-changed-05dc355c-bbf9-44a2-80dc-4a730699b8f8. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 676.661377] env[62814]: DEBUG oslo_concurrency.lockutils [req-e1db582f-f8d2-4cf0-a33c-46704589f271 req-30ca30be-06dc-4810-95c9-73fb075eb696 service nova] Acquiring lock "refresh_cache-98d134b4-b4ca-4247-a638-ad5c24a694e5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.661686] env[62814]: DEBUG oslo_concurrency.lockutils [req-e1db582f-f8d2-4cf0-a33c-46704589f271 req-30ca30be-06dc-4810-95c9-73fb075eb696 service nova] Acquired lock "refresh_cache-98d134b4-b4ca-4247-a638-ad5c24a694e5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.661957] env[62814]: DEBUG nova.network.neutron [req-e1db582f-f8d2-4cf0-a33c-46704589f271 req-30ca30be-06dc-4810-95c9-73fb075eb696 service nova] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Refreshing network info cache for port 05dc355c-bbf9-44a2-80dc-4a730699b8f8 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 676.672359] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Created folder: Project (eadee196b6264ba987043438ccdce5f8) in parent group-v845547. [ 676.672612] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Creating folder: Instances. Parent ref: group-v845602. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 676.672887] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-18cf898d-a39a-447b-9089-91077cc844ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.697336] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Created folder: Instances in parent group-v845602. [ 676.699847] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 676.699847] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 676.699847] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-30f9f785-23f5-45e1-8c63-b7185bca2ab4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.726979] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 676.726979] env[62814]: value = "task-4293345" [ 676.726979] env[62814]: _type = "Task" [ 676.726979] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.738598] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293345, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.821672] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293342, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.227649] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7523d2-6d2a-4718-a655-6d03fbd38342 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.246028] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc38759c-23a3-40ff-9105-b2c339247629 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.249979] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293345, 'name': CreateVM_Task, 'duration_secs': 0.515296} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.250189] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 677.251266] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.251424] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 677.251728] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 677.252255] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2ca10b1-d62c-4053-8bba-0f9623e82232 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.292048] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Successfully updated port: 04ddd249-262e-42ae-95d7-a32509b50e6d {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 677.294752] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe6dac9-29c4-45f9-bf21-184c623b6558 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.301027] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 677.301027] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52affe81-261b-04e8-59af-7f60e68a0bf2" [ 677.301027] env[62814]: _type = "Task" [ 677.301027] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.325540] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af22bf9-0fee-4e63-8fc0-7c403268c51a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.338695] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52affe81-261b-04e8-59af-7f60e68a0bf2, 'name': SearchDatastore_Task, 'duration_secs': 0.016586} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.339967] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 677.340304] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 677.340836] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.341032] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 677.341295] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 677.341935] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d670986d-26b6-4602-9834-c30818c6c27a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.355484] env[62814]: DEBUG nova.compute.provider_tree [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.360220] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293342, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.6334} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.360220] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98d134b4-b4ca-4247-a638-ad5c24a694e5/98d134b4-b4ca-4247-a638-ad5c24a694e5.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 677.360220] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 677.360220] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-77dee060-6734-4272-a37c-e40ad3bc391a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.368227] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 677.368227] env[62814]: value = "task-4293346" [ 677.368227] env[62814]: _type = "Task" [ 677.368227] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.368227] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 677.368696] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 677.369543] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35587961-ea01-4dc6-8844-78e01102f1e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.384120] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293346, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.384120] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 677.384120] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5283a757-4b8e-9b39-5d33-76d65c06f3cf" [ 677.384120] env[62814]: _type = "Task" [ 677.384120] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.396231] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5283a757-4b8e-9b39-5d33-76d65c06f3cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.467666] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cf621-8160-4031-40f2-0c893e2ec830/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 677.468873] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef7af29-fb1c-460a-8fd3-a90e97e4ec64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.475840] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cf621-8160-4031-40f2-0c893e2ec830/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 677.475937] env[62814]: ERROR oslo_vmware.rw_handles [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cf621-8160-4031-40f2-0c893e2ec830/disk-0.vmdk due to incomplete transfer. [ 677.476172] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a274c360-2008-4ce2-b2d7-3bfed9f3454a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.485944] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/520cf621-8160-4031-40f2-0c893e2ec830/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 677.486052] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Uploaded image e7328b9e-7c4e-41a4-ade1-d439df2073ed to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 677.487995] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 677.490491] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-05782625-d6a6-49a0-82a5-6d4b049132dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.497775] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 677.497775] env[62814]: value = "task-4293347" [ 677.497775] env[62814]: _type = "Task" [ 677.497775] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.508628] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293347, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.630057] env[62814]: DEBUG nova.network.neutron [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Updating instance_info_cache with network_info: [{"id": "a2e379d6-9cd3-4a22-85c2-72628174b4a6", "address": "fa:16:3e:3d:52:e0", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa2e379d6-9c", "ovs_interfaceid": "a2e379d6-9cd3-4a22-85c2-72628174b4a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.802159] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "refresh_cache-8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.802159] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "refresh_cache-8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 677.802159] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.861249] env[62814]: DEBUG nova.scheduler.client.report [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 677.868763] env[62814]: DEBUG nova.network.neutron [req-e1db582f-f8d2-4cf0-a33c-46704589f271 req-30ca30be-06dc-4810-95c9-73fb075eb696 service nova] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Updated VIF entry in instance network info cache for port 05dc355c-bbf9-44a2-80dc-4a730699b8f8. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 677.869112] env[62814]: DEBUG nova.network.neutron [req-e1db582f-f8d2-4cf0-a33c-46704589f271 req-30ca30be-06dc-4810-95c9-73fb075eb696 service nova] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Updating instance_info_cache with network_info: [{"id": "05dc355c-bbf9-44a2-80dc-4a730699b8f8", "address": "fa:16:3e:7c:07:98", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.136", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05dc355c-bb", "ovs_interfaceid": "05dc355c-bbf9-44a2-80dc-4a730699b8f8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.881418] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293346, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073697} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.881692] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 677.882509] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad47a49f-bd53-4946-85bf-b0fcc700c8fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.914747] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] 98d134b4-b4ca-4247-a638-ad5c24a694e5/98d134b4-b4ca-4247-a638-ad5c24a694e5.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 677.920333] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-821cb4c8-b1ff-44a3-adc4-4f0f868dc6a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.935581] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5283a757-4b8e-9b39-5d33-76d65c06f3cf, 'name': SearchDatastore_Task, 'duration_secs': 0.01395} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.935581] env[62814]: DEBUG nova.network.neutron [-] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.939604] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0aa53e8e-db26-46ae-a07d-7692bc34b20f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.945282] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 677.945282] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e26d73-bb1e-5993-4cbc-e36e0ad83572" [ 677.945282] env[62814]: _type = "Task" [ 677.945282] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.948018] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 677.948018] env[62814]: value = "task-4293348" [ 677.948018] env[62814]: _type = "Task" [ 677.948018] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.970028] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293348, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.975383] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e26d73-bb1e-5993-4cbc-e36e0ad83572, 'name': SearchDatastore_Task, 'duration_secs': 0.012471} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.975383] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 677.975634] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a36ad785-2f33-4dbc-bc82-ab4a35020b0f/a36ad785-2f33-4dbc-bc82-ab4a35020b0f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 677.975883] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7af1c23d-7643-4616-baa6-11ff2ddee3a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.984863] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 677.984863] env[62814]: value = "task-4293349" [ 677.984863] env[62814]: _type = "Task" [ 677.984863] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.994709] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293349, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.008974] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293347, 'name': Destroy_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.132850] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "refresh_cache-d9242042-6209-4b04-bf00-00dd04d9d6a0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.375230] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 4.276s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 678.376083] env[62814]: INFO nova.compute.manager [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Migrating [ 678.376083] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.376083] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 678.378176] env[62814]: DEBUG oslo_concurrency.lockutils [req-e1db582f-f8d2-4cf0-a33c-46704589f271 req-30ca30be-06dc-4810-95c9-73fb075eb696 service nova] Releasing lock "refresh_cache-98d134b4-b4ca-4247-a638-ad5c24a694e5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.379012] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.193s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 678.379366] env[62814]: DEBUG nova.objects.instance [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Lazy-loading 'resources' on Instance uuid 0ebb226a-40a3-40cd-8492-ccce87baf0e6 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 678.384833] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.441675] env[62814]: INFO nova.compute.manager [-] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Took 2.24 seconds to deallocate network for instance. [ 678.469698] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293348, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.476507] env[62814]: DEBUG nova.compute.manager [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Received event network-changed-fb397a6f-5617-48a5-b5f6-30d9be90c5d4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 678.478026] env[62814]: DEBUG nova.compute.manager [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Refreshing instance network info cache due to event network-changed-fb397a6f-5617-48a5-b5f6-30d9be90c5d4. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 678.478026] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] Acquiring lock "refresh_cache-a36ad785-2f33-4dbc-bc82-ab4a35020b0f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.478026] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] Acquired lock "refresh_cache-a36ad785-2f33-4dbc-bc82-ab4a35020b0f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 678.478026] env[62814]: DEBUG nova.network.neutron [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Refreshing network info cache for port fb397a6f-5617-48a5-b5f6-30d9be90c5d4 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 678.506382] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293349, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.526793] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293347, 'name': Destroy_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.836012] env[62814]: DEBUG nova.network.neutron [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Updating instance_info_cache with network_info: [{"id": "04ddd249-262e-42ae-95d7-a32509b50e6d", "address": "fa:16:3e:22:40:4b", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04ddd249-26", "ovs_interfaceid": "04ddd249-262e-42ae-95d7-a32509b50e6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.880042] env[62814]: INFO nova.compute.rpcapi [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 678.880473] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.948635] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 678.970601] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293348, 'name': ReconfigVM_Task, 'duration_secs': 0.649487} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.971349] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Reconfigured VM instance instance-00000011 to attach disk [datastore2] 98d134b4-b4ca-4247-a638-ad5c24a694e5/98d134b4-b4ca-4247-a638-ad5c24a694e5.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 678.972100] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce48c71b-44a4-436b-8729-40d2b92b7ca7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.980669] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 678.980669] env[62814]: value = "task-4293350" [ 678.980669] env[62814]: _type = "Task" [ 678.980669] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.993902] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293350, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.002288] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293349, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556144} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.002545] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a36ad785-2f33-4dbc-bc82-ab4a35020b0f/a36ad785-2f33-4dbc-bc82-ab4a35020b0f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 679.002747] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 679.002985] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-35d5ad68-12c0-47f9-857c-2a1e0a9d2582 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.011518] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 679.011518] env[62814]: value = "task-4293351" [ 679.011518] env[62814]: _type = "Task" [ 679.011518] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.021455] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293347, 'name': Destroy_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.027327] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293351, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.250657] env[62814]: DEBUG nova.compute.manager [req-29bfd804-7486-4c9c-8050-5a1175829d97 req-b2f4a4a1-9cb1-4124-9131-b2319d1c2274 service nova] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Received event network-vif-deleted-4615b448-a51a-4bfb-85fb-fd8f118e479d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 679.338372] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "refresh_cache-8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 679.338785] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Instance network_info: |[{"id": "04ddd249-262e-42ae-95d7-a32509b50e6d", "address": "fa:16:3e:22:40:4b", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04ddd249-26", "ovs_interfaceid": "04ddd249-262e-42ae-95d7-a32509b50e6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 679.339261] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:40:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '04ddd249-262e-42ae-95d7-a32509b50e6d', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 679.349279] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 679.353037] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 679.353518] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5c61267-335b-4039-a8a9-2d37f8af664b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.379865] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 679.379865] env[62814]: value = "task-4293352" [ 679.379865] env[62814]: _type = "Task" [ 679.379865] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.393255] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293352, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.402698] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.402893] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.403169] env[62814]: DEBUG nova.network.neutron [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 679.449422] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5750a66f-8039-4ece-9bbc-b439430265e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.460111] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841434c8-0f2b-42bd-88e3-b76dce2453f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.501284] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0173b9-8baf-46d7-94b6-1700b2c13a8f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.511213] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293350, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.517803] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d714409e-c56f-4cd0-9ab7-1bd2b63fcd55 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.531552] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293347, 'name': Destroy_Task, 'duration_secs': 1.814292} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.541548] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Destroyed the VM [ 679.541912] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 679.542495] env[62814]: DEBUG nova.compute.provider_tree [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.543879] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293351, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083847} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.544101] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-716a75c7-90e7-4d69-9f9b-13100de7b518 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.545902] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 679.546955] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ec5b090-9700-4d65-b46a-8ccd7cc0f71b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.577146] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] a36ad785-2f33-4dbc-bc82-ab4a35020b0f/a36ad785-2f33-4dbc-bc82-ab4a35020b0f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 679.583197] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4bef8aee-3480-474f-9c91-746b63690f7f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.598688] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 679.598688] env[62814]: value = "task-4293353" [ 679.598688] env[62814]: _type = "Task" [ 679.598688] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.607151] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 679.607151] env[62814]: value = "task-4293354" [ 679.607151] env[62814]: _type = "Task" [ 679.607151] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.614991] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293353, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.621765] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293354, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.691464] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 679.691677] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a0c17ac4-711e-4c86-8f41-b6b2de41059f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.708000] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 679.708000] env[62814]: value = "task-4293355" [ 679.708000] env[62814]: _type = "Task" [ 679.708000] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.725402] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.763837] env[62814]: DEBUG nova.network.neutron [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Updated VIF entry in instance network info cache for port fb397a6f-5617-48a5-b5f6-30d9be90c5d4. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 679.763837] env[62814]: DEBUG nova.network.neutron [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Updating instance_info_cache with network_info: [{"id": "fb397a6f-5617-48a5-b5f6-30d9be90c5d4", "address": "fa:16:3e:08:4a:69", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb397a6f-56", "ovs_interfaceid": "fb397a6f-5617-48a5-b5f6-30d9be90c5d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.894197] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293352, 'name': CreateVM_Task, 'duration_secs': 0.410514} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.894428] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 679.895306] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.895472] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.896346] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 679.896346] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d7a2820-780c-44ae-93bf-a103fd585672 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.902365] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 679.902365] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52df7e8a-6eca-3c97-5ba0-40dd3c9912d7" [ 679.902365] env[62814]: _type = "Task" [ 679.902365] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.921348] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52df7e8a-6eca-3c97-5ba0-40dd3c9912d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.014297] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293350, 'name': Rename_Task, 'duration_secs': 0.68187} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.014614] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 680.014821] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b0e9a26-8ce4-43e7-9449-4d1a2d4adf17 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.023663] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 680.023663] env[62814]: value = "task-4293356" [ 680.023663] env[62814]: _type = "Task" [ 680.023663] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.036224] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293356, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.048516] env[62814]: DEBUG nova.scheduler.client.report [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 680.112938] env[62814]: DEBUG oslo_vmware.api [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293353, 'name': RemoveSnapshot_Task, 'duration_secs': 0.552914} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.117209] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 680.117865] env[62814]: INFO nova.compute.manager [None req-dc9588aa-7b0d-4f8d-9bb2-28b1a8640ab2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Took 17.29 seconds to snapshot the instance on the hypervisor. [ 680.126349] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293354, 'name': ReconfigVM_Task, 'duration_secs': 0.336028} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.126876] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Reconfigured VM instance instance-00000012 to attach disk [datastore2] a36ad785-2f33-4dbc-bc82-ab4a35020b0f/a36ad785-2f33-4dbc-bc82-ab4a35020b0f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 680.127311] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f4eb9cbe-c96f-454c-9e70-58d84994b6a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.137108] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 680.137108] env[62814]: value = "task-4293357" [ 680.137108] env[62814]: _type = "Task" [ 680.137108] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.146085] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293357, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.222420] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293355, 'name': PowerOffVM_Task, 'duration_secs': 0.209323} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.222674] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 680.224401] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e22498-20cb-452f-9614-3cad679e04c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.245675] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae99628-76c0-413f-96f2-3411d1fff9ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.267173] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] Releasing lock "refresh_cache-a36ad785-2f33-4dbc-bc82-ab4a35020b0f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.267459] env[62814]: DEBUG nova.compute.manager [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Received event network-vif-plugged-04ddd249-262e-42ae-95d7-a32509b50e6d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 680.267655] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] Acquiring lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.267862] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] Lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.268037] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] Lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.268208] env[62814]: DEBUG nova.compute.manager [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] No waiting events found dispatching network-vif-plugged-04ddd249-262e-42ae-95d7-a32509b50e6d {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 680.268406] env[62814]: WARNING nova.compute.manager [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Received unexpected event network-vif-plugged-04ddd249-262e-42ae-95d7-a32509b50e6d for instance with vm_state building and task_state spawning. [ 680.268638] env[62814]: DEBUG nova.compute.manager [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Received event network-changed-04ddd249-262e-42ae-95d7-a32509b50e6d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 680.268830] env[62814]: DEBUG nova.compute.manager [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Refreshing instance network info cache due to event network-changed-04ddd249-262e-42ae-95d7-a32509b50e6d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 680.269078] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] Acquiring lock "refresh_cache-8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.269259] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] Acquired lock "refresh_cache-8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.269450] env[62814]: DEBUG nova.network.neutron [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Refreshing network info cache for port 04ddd249-262e-42ae-95d7-a32509b50e6d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 680.287052] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 680.287812] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c252875-6f45-4d9d-a813-880368b37945 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.301897] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 680.301897] env[62814]: value = "task-4293358" [ 680.301897] env[62814]: _type = "Task" [ 680.301897] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.310941] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293358, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.367669] env[62814]: DEBUG nova.network.neutron [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance_info_cache with network_info: [{"id": "827a94b4-5864-4060-bfb5-d0e9d2281332", "address": "fa:16:3e:a9:8f:95", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.43", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap827a94b4-58", "ovs_interfaceid": "827a94b4-5864-4060-bfb5-d0e9d2281332", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.414728] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52df7e8a-6eca-3c97-5ba0-40dd3c9912d7, 'name': SearchDatastore_Task, 'duration_secs': 0.015522} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.415041] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.416294] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 680.416294] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.416294] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.416294] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 680.416294] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bd4bfc1f-7bd2-4358-98bc-b68b562f8002 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.427976] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 680.427976] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 680.428302] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8efe79bf-3540-473b-b618-44a5b5a5f270 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.435679] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 680.435679] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f6ac8d-c02d-e1cd-f161-4d6263c4886c" [ 680.435679] env[62814]: _type = "Task" [ 680.435679] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.446440] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f6ac8d-c02d-e1cd-f161-4d6263c4886c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.545301] env[62814]: DEBUG oslo_vmware.api [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293356, 'name': PowerOnVM_Task, 'duration_secs': 0.488653} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.545735] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 680.545882] env[62814]: INFO nova.compute.manager [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Took 11.20 seconds to spawn the instance on the hypervisor. [ 680.546167] env[62814]: DEBUG nova.compute.manager [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 680.548899] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9162e607-2a06-4154-9e9e-29693b228bd2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.553533] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.174s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.557825] env[62814]: DEBUG oslo_concurrency.lockutils [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.386s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.557825] env[62814]: DEBUG nova.objects.instance [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lazy-loading 'resources' on Instance uuid 66db614c-d101-409d-a5c3-9de38e5a6eff {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 680.585763] env[62814]: INFO nova.scheduler.client.report [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Deleted allocations for instance 0ebb226a-40a3-40cd-8492-ccce87baf0e6 [ 680.650715] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293357, 'name': Rename_Task, 'duration_secs': 0.169321} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.650715] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 680.650715] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-75e709bb-63b3-4880-b200-cd5522568cc3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.658122] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 680.658122] env[62814]: value = "task-4293359" [ 680.658122] env[62814]: _type = "Task" [ 680.658122] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.671074] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293359, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.817038] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 680.817038] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 680.817602] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.873825] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.954540] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f6ac8d-c02d-e1cd-f161-4d6263c4886c, 'name': SearchDatastore_Task, 'duration_secs': 0.011844} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.958323] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9fff615-e39b-4be0-807f-566e5a631f72 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.969087] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 680.969087] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc84d0-a17d-6c13-6619-438430dc1869" [ 680.969087] env[62814]: _type = "Task" [ 680.969087] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.989866] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc84d0-a17d-6c13-6619-438430dc1869, 'name': SearchDatastore_Task, 'duration_secs': 0.013737} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.990175] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.990569] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e/8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 680.992103] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.992103] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 680.992103] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5713efe5-51fc-48c7-bc89-72901541e5b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.995759] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec477f19-2642-4169-bd97-1c2a3ca5dd25 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.009012] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 681.009012] env[62814]: value = "task-4293360" [ 681.009012] env[62814]: _type = "Task" [ 681.009012] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.010325] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 681.013524] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 681.015018] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b145a89-c959-46c4-a35f-e64a19cb513b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.030402] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 681.030402] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cac234-8ac7-2e33-5038-7086ace51578" [ 681.030402] env[62814]: _type = "Task" [ 681.030402] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.030707] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293360, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.046249] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cac234-8ac7-2e33-5038-7086ace51578, 'name': SearchDatastore_Task, 'duration_secs': 0.01354} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.050323] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80a243a9-1b7e-47f1-ac1f-d15ef2e15f0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.058889] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 681.058889] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a0cbd3-2aa1-a39b-92ee-ea6a3a202019" [ 681.058889] env[62814]: _type = "Task" [ 681.058889] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.085606] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a0cbd3-2aa1-a39b-92ee-ea6a3a202019, 'name': SearchDatastore_Task, 'duration_secs': 0.011338} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.088697] env[62814]: INFO nova.compute.manager [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Took 37.20 seconds to build instance. [ 681.088697] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.088697] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d9242042-6209-4b04-bf00-00dd04d9d6a0/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. {{(pid=62814) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 681.088697] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-856b2147-312f-403f-b991-a1039f140ba3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.098741] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 681.098741] env[62814]: value = "task-4293361" [ 681.098741] env[62814]: _type = "Task" [ 681.098741] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.099235] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d0b0afe-355e-4297-8480-4cafddcc55ad tempest-DeleteServersAdminTestJSON-80280349 tempest-DeleteServersAdminTestJSON-80280349-project-admin] Lock "0ebb226a-40a3-40cd-8492-ccce87baf0e6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.021s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 681.110668] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293361, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.175138] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293359, 'name': PowerOnVM_Task, 'duration_secs': 0.468365} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.175436] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 681.175701] env[62814]: INFO nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Took 7.88 seconds to spawn the instance on the hypervisor. [ 681.175937] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 681.176868] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-884bbdb9-b335-46a0-813d-9ff51a3c9a00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.367215] env[62814]: DEBUG nova.network.neutron [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Updated VIF entry in instance network info cache for port 04ddd249-262e-42ae-95d7-a32509b50e6d. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 681.367215] env[62814]: DEBUG nova.network.neutron [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Updating instance_info_cache with network_info: [{"id": "04ddd249-262e-42ae-95d7-a32509b50e6d", "address": "fa:16:3e:22:40:4b", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04ddd249-26", "ovs_interfaceid": "04ddd249-262e-42ae-95d7-a32509b50e6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.525236] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293360, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.539919] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "62bc755d-4f96-4486-884b-0d0c337267aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.540691] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "62bc755d-4f96-4486-884b-0d0c337267aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.541097] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "62bc755d-4f96-4486-884b-0d0c337267aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.541449] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "62bc755d-4f96-4486-884b-0d0c337267aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.541729] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "62bc755d-4f96-4486-884b-0d0c337267aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 681.548691] env[62814]: INFO nova.compute.manager [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Terminating instance [ 681.589670] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c3f2457-b193-4465-a463-4f3d338bad72 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "98d134b4-b4ca-4247-a638-ad5c24a694e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.717s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 681.621271] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293361, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.713074] env[62814]: INFO nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Took 33.46 seconds to build instance. [ 681.810861] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3abe579-260e-4e7f-b516-6ff4f2c7df04 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.827090] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11926b67-2029-40a3-a3b1-4c5722a54476 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.900280] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a808811-7bb9-41fb-a0e8-a523041fb46f req-328b60ef-ca8d-4091-b806-47fa77c3c773 service nova] Releasing lock "refresh_cache-8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.903874] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2692d660-992e-4213-b4d4-b96f05fb9563 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.920719] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30cd7d3b-6365-4402-9d4c-abc7c9cdfafe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.944597] env[62814]: DEBUG nova.compute.provider_tree [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.028174] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293360, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.608118} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.028460] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e/8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 682.028802] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 682.029180] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aff35c85-d779-43e3-a6da-8da8fc029db5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.040881] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 682.040881] env[62814]: value = "task-4293362" [ 682.040881] env[62814]: _type = "Task" [ 682.040881] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.057806] env[62814]: DEBUG nova.compute.manager [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 682.057806] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 682.058051] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293362, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.058870] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de48de29-5187-45c7-9329-a54383a38ad4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.068153] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 682.068153] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4507d227-1e4f-440f-9f52-865185b6a32a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.093112] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 682.113759] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293361, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.868394} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.114071] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d9242042-6209-4b04-bf00-00dd04d9d6a0/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. [ 682.114926] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae8c670-c2ab-4aaa-a722-72f531e7319b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.145872] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] d9242042-6209-4b04-bf00-00dd04d9d6a0/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 682.149020] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78d86977-3f01-4ec2-aa4d-d9a4677b847e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.171285] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 682.171285] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 682.171285] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleting the datastore file [datastore2] 62bc755d-4f96-4486-884b-0d0c337267aa {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 682.171285] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-73e78f6d-7b50-4ac7-b5c0-c7e0bc650c73 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.179009] env[62814]: DEBUG oslo_vmware.api [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 682.179009] env[62814]: value = "task-4293364" [ 682.179009] env[62814]: _type = "Task" [ 682.179009] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.184946] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 682.184946] env[62814]: value = "task-4293365" [ 682.184946] env[62814]: _type = "Task" [ 682.184946] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.192201] env[62814]: DEBUG oslo_vmware.api [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293364, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.197788] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293365, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.215133] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.978s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 682.413157] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab22e41-36da-4433-afda-9daf04ff4a63 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.435721] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance '6a592192-1b41-4be2-84a6-c3b76a4e5643' progress to 0 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 682.450157] env[62814]: DEBUG nova.scheduler.client.report [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 682.556743] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293362, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100331} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.557053] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 682.557856] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242b4d2b-7ee5-4f71-b28d-130aed9fc7a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.584721] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Reconfiguring VM instance instance-00000013 to attach disk [datastore2] 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e/8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 682.585086] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-655722d9-0b06-4c81-87bb-c2cb4422f913 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.614261] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 682.614261] env[62814]: value = "task-4293366" [ 682.614261] env[62814]: _type = "Task" [ 682.614261] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.625532] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293366, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.639053] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 682.693552] env[62814]: DEBUG oslo_vmware.api [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293364, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.699976] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293365, 'name': ReconfigVM_Task, 'duration_secs': 0.310133} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.700275] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Reconfigured VM instance instance-00000010 to attach disk [datastore2] d9242042-6209-4b04-bf00-00dd04d9d6a0/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 682.700275] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e44a3f7e-e8e4-474c-88fe-21f651b392a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.728438] env[62814]: DEBUG nova.compute.manager [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 682.737813] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd7f22ab-dd84-4571-b5be-1a11508abcab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.762111] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 682.762111] env[62814]: value = "task-4293367" [ 682.762111] env[62814]: _type = "Task" [ 682.762111] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.774721] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293367, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.942541] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 682.942985] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0c10fd03-277f-4dc0-bf88-84816b5c3baf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.954967] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 682.954967] env[62814]: value = "task-4293368" [ 682.954967] env[62814]: _type = "Task" [ 682.954967] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.972337] env[62814]: DEBUG oslo_concurrency.lockutils [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.414s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 682.972850] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.282s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 682.974547] env[62814]: INFO nova.compute.claims [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 683.008350] env[62814]: INFO nova.scheduler.client.report [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Deleted allocations for instance 66db614c-d101-409d-a5c3-9de38e5a6eff [ 683.131664] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293366, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.194941] env[62814]: DEBUG oslo_vmware.api [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293364, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.532179} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.194941] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 683.195195] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 683.195989] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.195989] env[62814]: INFO nova.compute.manager [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Took 1.14 seconds to destroy the instance on the hypervisor. [ 683.195989] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 683.195989] env[62814]: DEBUG nova.compute.manager [-] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 683.195989] env[62814]: DEBUG nova.network.neutron [-] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.259326] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 683.259326] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 683.268245] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 683.276075] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293367, 'name': ReconfigVM_Task, 'duration_secs': 0.168741} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.278516] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 683.279161] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c9d8213c-44a9-4ab4-b6f4-aedddde31c8e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.302341] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 683.302341] env[62814]: value = "task-4293369" [ 683.302341] env[62814]: _type = "Task" [ 683.302341] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.316799] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293369, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.384786] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8112776-5d2b-46fa-baeb-3fcf37eeca5c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.395714] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4677daeb-a706-438c-b148-d8eace2cd9b3 tempest-ServersAdminNegativeTestJSON-1819159048 tempest-ServersAdminNegativeTestJSON-1819159048-project-admin] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Suspending the VM {{(pid=62814) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 683.398432] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-36730c60-d909-4a7a-97cc-ef66e78bbfbc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.404154] env[62814]: DEBUG oslo_vmware.api [None req-4677daeb-a706-438c-b148-d8eace2cd9b3 tempest-ServersAdminNegativeTestJSON-1819159048 tempest-ServersAdminNegativeTestJSON-1819159048-project-admin] Waiting for the task: (returnval){ [ 683.404154] env[62814]: value = "task-4293370" [ 683.404154] env[62814]: _type = "Task" [ 683.404154] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.414584] env[62814]: DEBUG oslo_vmware.api [None req-4677daeb-a706-438c-b148-d8eace2cd9b3 tempest-ServersAdminNegativeTestJSON-1819159048 tempest-ServersAdminNegativeTestJSON-1819159048-project-admin] Task: {'id': task-4293370, 'name': SuspendVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.465891] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293368, 'name': PowerOffVM_Task, 'duration_secs': 0.370208} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.466371] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 683.466616] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance '6a592192-1b41-4be2-84a6-c3b76a4e5643' progress to 17 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 683.518081] env[62814]: DEBUG oslo_concurrency.lockutils [None req-034025e9-8420-48a7-a014-c8db91e32217 tempest-TenantUsagesTestJSON-551476932 tempest-TenantUsagesTestJSON-551476932-project-member] Lock "66db614c-d101-409d-a5c3-9de38e5a6eff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.335s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.629776] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293366, 'name': ReconfigVM_Task, 'duration_secs': 0.646489} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.630176] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Reconfigured VM instance instance-00000013 to attach disk [datastore2] 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e/8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 683.631082] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5c0a9a63-d2db-4f27-be3a-01dc14aaa4f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.641617] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 683.641617] env[62814]: value = "task-4293371" [ 683.641617] env[62814]: _type = "Task" [ 683.641617] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.651105] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293371, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.815149] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293369, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.916730] env[62814]: DEBUG oslo_vmware.api [None req-4677daeb-a706-438c-b148-d8eace2cd9b3 tempest-ServersAdminNegativeTestJSON-1819159048 tempest-ServersAdminNegativeTestJSON-1819159048-project-admin] Task: {'id': task-4293370, 'name': SuspendVM_Task} progress is 62%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.977592] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 683.978171] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 683.978171] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 683.978300] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 683.978382] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 683.981038] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 683.981038] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 683.981038] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 683.981038] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 683.981038] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 683.981395] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 683.991017] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-511272c1-0fde-4da8-8809-091639a6c5fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.012639] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 684.012639] env[62814]: value = "task-4293372" [ 684.012639] env[62814]: _type = "Task" [ 684.012639] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.023848] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293372, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.092335] env[62814]: DEBUG nova.network.neutron [-] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.153351] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293371, 'name': Rename_Task, 'duration_secs': 0.234896} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.153646] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 684.153920] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a7f01f37-38b4-4129-acd0-9f0cc0e4889a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.164194] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 684.164194] env[62814]: value = "task-4293373" [ 684.164194] env[62814]: _type = "Task" [ 684.164194] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.171488] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293373, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.316175] env[62814]: DEBUG oslo_vmware.api [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293369, 'name': PowerOnVM_Task, 'duration_secs': 0.547929} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.321264] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 684.324139] env[62814]: DEBUG nova.compute.manager [None req-9b2348cd-c914-4e84-9fec-ac73515a1ef7 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 684.324961] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d42d7ae-8219-43ad-8a3e-3146ec2e5c5e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.416139] env[62814]: DEBUG oslo_vmware.api [None req-4677daeb-a706-438c-b148-d8eace2cd9b3 tempest-ServersAdminNegativeTestJSON-1819159048 tempest-ServersAdminNegativeTestJSON-1819159048-project-admin] Task: {'id': task-4293370, 'name': SuspendVM_Task, 'duration_secs': 0.681181} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.419178] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4677daeb-a706-438c-b148-d8eace2cd9b3 tempest-ServersAdminNegativeTestJSON-1819159048 tempest-ServersAdminNegativeTestJSON-1819159048-project-admin] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Suspended the VM {{(pid=62814) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 684.419178] env[62814]: DEBUG nova.compute.manager [None req-4677daeb-a706-438c-b148-d8eace2cd9b3 tempest-ServersAdminNegativeTestJSON-1819159048 tempest-ServersAdminNegativeTestJSON-1819159048-project-admin] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 684.420159] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27453bed-5aab-4956-aec2-25969aac1bf8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.476296] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.476886] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.529212] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293372, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.596231] env[62814]: INFO nova.compute.manager [-] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Took 1.40 seconds to deallocate network for instance. [ 684.610018] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14cdf05c-99a3-416a-89c4-07172ad20923 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.619606] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d11cfff-95f3-432a-aff9-cd0256b43d96 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.626157] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "9343301c-dfe9-41b0-b4a0-067af544d297" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.626460] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.659020] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8825147-2757-4ecc-aa44-fb4383b1b36e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.670090] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5195b09a-772c-44f0-9a6c-6ede203e47af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.678363] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293373, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.690438] env[62814]: DEBUG nova.compute.provider_tree [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.739175] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "56f9e03c-1157-442f-8add-156627a6ec1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 684.739430] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "56f9e03c-1157-442f-8add-156627a6ec1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.861116] env[62814]: DEBUG nova.compute.manager [req-d0e7ed50-a977-4c27-83a3-6aff4a8e9861 req-b46f90ea-62be-496a-b690-f02198ed8fe7 service nova] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Received event network-vif-deleted-559634c5-106d-4de8-b7b2-596fd81a0c30 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 685.025483] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293372, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.106587] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 685.175203] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293373, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.196162] env[62814]: DEBUG nova.scheduler.client.report [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 685.528774] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293372, 'name': ReconfigVM_Task, 'duration_secs': 1.019371} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.529104] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance '6a592192-1b41-4be2-84a6-c3b76a4e5643' progress to 33 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 685.675206] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293373, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.704566] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.732s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 685.705220] env[62814]: DEBUG nova.compute.manager [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 685.712090] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.843s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 685.712220] env[62814]: DEBUG nova.objects.instance [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lazy-loading 'resources' on Instance uuid 591e4a92-7c03-4d70-885f-49a66032c97e {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 685.821485] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquiring lock "1a16bd56-0992-4bec-bd3f-2836dfe68579" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 685.821485] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "1a16bd56-0992-4bec-bd3f-2836dfe68579" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.035607] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 686.035922] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.036061] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 686.036250] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.036404] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 686.036547] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 686.036752] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 686.036911] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 686.037295] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 686.037433] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 686.037953] env[62814]: DEBUG nova.virt.hardware [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 686.043648] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Reconfiguring VM instance instance-00000008 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 686.043965] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f51253e-1129-4b72-9cae-da8d85aa2b5d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.065828] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 686.065828] env[62814]: value = "task-4293374" [ 686.065828] env[62814]: _type = "Task" [ 686.065828] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.081745] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293374, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.177271] env[62814]: DEBUG oslo_vmware.api [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293373, 'name': PowerOnVM_Task, 'duration_secs': 1.605423} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.177687] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 686.178359] env[62814]: INFO nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Took 10.04 seconds to spawn the instance on the hypervisor. [ 686.178359] env[62814]: DEBUG nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 686.179040] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b75b3ea-485d-4634-81f7-d521e071f0e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.217416] env[62814]: DEBUG nova.compute.utils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 686.221717] env[62814]: DEBUG nova.compute.manager [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 686.221964] env[62814]: DEBUG nova.network.neutron [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 686.301546] env[62814]: DEBUG nova.policy [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51348be2493643fb9e5b44f27e5a7a94', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '513135674ee446d19fa8c667a47138a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 686.586897] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293374, 'name': ReconfigVM_Task, 'duration_secs': 0.268247} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.587378] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Reconfigured VM instance instance-00000008 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 686.588739] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a312b8-ae65-4dc4-9548-364bcbd2a490 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.618502] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Reconfiguring VM instance instance-00000008 to attach disk [datastore2] 6a592192-1b41-4be2-84a6-c3b76a4e5643/6a592192-1b41-4be2-84a6-c3b76a4e5643.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 686.621680] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6ead4d9-f0f9-42b8-9fce-2c2079ccc250 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.646791] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 686.646791] env[62814]: value = "task-4293375" [ 686.646791] env[62814]: _type = "Task" [ 686.646791] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.656731] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293375, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.703834] env[62814]: INFO nova.compute.manager [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Took 38.38 seconds to build instance. [ 686.722239] env[62814]: DEBUG nova.compute.manager [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 686.941367] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea42c2d5-4cc8-4485-8f64-041417da9eb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.951550] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f49d171a-99fc-47c4-a723-05f42cf8a881 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.991561] env[62814]: DEBUG nova.network.neutron [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Successfully created port: a00402c1-ebc9-40c0-93b2-26b6fbab4a68 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.995041] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0c9997-a0e8-42c4-8d9f-d637a0541fd6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.003773] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c8487f-af75-41c7-9806-6ba3b274a93d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.018702] env[62814]: DEBUG nova.compute.provider_tree [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.157673] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293375, 'name': ReconfigVM_Task, 'duration_secs': 0.45664} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.157962] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Reconfigured VM instance instance-00000008 to attach disk [datastore2] 6a592192-1b41-4be2-84a6-c3b76a4e5643/6a592192-1b41-4be2-84a6-c3b76a4e5643.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 687.158248] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance '6a592192-1b41-4be2-84a6-c3b76a4e5643' progress to 50 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 687.210449] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34b53e5c-b3bf-4361-876d-771fa06a03b6 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.900s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 687.522803] env[62814]: DEBUG nova.scheduler.client.report [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 687.665461] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a37348-04c4-4bf0-93a9-39a469a99b0b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.689758] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-449c5a6d-6647-42b2-851c-495ecae63f7d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.708935] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance '6a592192-1b41-4be2-84a6-c3b76a4e5643' progress to 67 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 687.715385] env[62814]: DEBUG nova.compute.manager [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 687.741348] env[62814]: DEBUG nova.compute.manager [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 687.794873] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 687.795264] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.795457] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 687.795666] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.795883] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 687.796097] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 687.796354] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 687.796553] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 687.796799] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 687.797007] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 687.797236] env[62814]: DEBUG nova.virt.hardware [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 687.798562] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b461c31-3a90-46e8-8ef7-66b3e8a307a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.810265] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c2e816-72f0-4961-9d57-0e757b3eba5b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.033049] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.321s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 688.035806] env[62814]: DEBUG oslo_concurrency.lockutils [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.125s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 688.036310] env[62814]: DEBUG nova.objects.instance [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lazy-loading 'resources' on Instance uuid f5504590-823d-4c37-8226-26684f7957a6 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 688.137849] env[62814]: INFO nova.scheduler.client.report [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Deleted allocations for instance 591e4a92-7c03-4d70-885f-49a66032c97e [ 688.246410] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 688.268934] env[62814]: DEBUG nova.network.neutron [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Port 827a94b4-5864-4060-bfb5-d0e9d2281332 binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 688.365910] env[62814]: INFO nova.compute.manager [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Rescuing [ 688.366211] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.366359] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 688.367786] env[62814]: DEBUG nova.network.neutron [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.656179] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3edfeba3-d326-410d-9916-4f9e470ba9b4 tempest-ServerDiagnosticsV248Test-1416942857 tempest-ServerDiagnosticsV248Test-1416942857-project-member] Lock "591e4a92-7c03-4d70-885f-49a66032c97e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.136s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.020793] env[62814]: DEBUG oslo_concurrency.lockutils [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.021538] env[62814]: DEBUG oslo_concurrency.lockutils [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.021538] env[62814]: DEBUG oslo_concurrency.lockutils [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.021993] env[62814]: DEBUG oslo_concurrency.lockutils [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.022048] env[62814]: DEBUG oslo_concurrency.lockutils [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.025251] env[62814]: INFO nova.compute.manager [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Terminating instance [ 689.049457] env[62814]: DEBUG nova.network.neutron [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Successfully updated port: a00402c1-ebc9-40c0-93b2-26b6fbab4a68 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 689.136039] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4cfa4d-e115-4106-a6f9-ebc93d1d4481 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.146643] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8da3c5b-aa3f-44c5-a0cf-56d76d2f2db6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.184865] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de74caee-a3da-410c-a8fc-97448a3e47e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.194912] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f08552-60a7-4666-8410-b21710141448 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.213722] env[62814]: DEBUG nova.compute.provider_tree [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.222928] env[62814]: DEBUG nova.network.neutron [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Updating instance_info_cache with network_info: [{"id": "22bf5740-3f2e-45ed-be98-ac62c70cfb3b", "address": "fa:16:3e:6e:e4:ef", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22bf5740-3f", "ovs_interfaceid": "22bf5740-3f2e-45ed-be98-ac62c70cfb3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.303767] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "6a592192-1b41-4be2-84a6-c3b76a4e5643-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.304227] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.304835] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.462906] env[62814]: DEBUG nova.compute.manager [req-9c6cad8c-defb-4e15-a330-f187ac121fcc req-2d2716c9-1cf1-42df-8162-dfbc5438412a service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Received event network-vif-plugged-a00402c1-ebc9-40c0-93b2-26b6fbab4a68 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 689.462906] env[62814]: DEBUG oslo_concurrency.lockutils [req-9c6cad8c-defb-4e15-a330-f187ac121fcc req-2d2716c9-1cf1-42df-8162-dfbc5438412a service nova] Acquiring lock "59d21ef7-df97-49ac-9329-4c18df6dd087-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.462906] env[62814]: DEBUG oslo_concurrency.lockutils [req-9c6cad8c-defb-4e15-a330-f187ac121fcc req-2d2716c9-1cf1-42df-8162-dfbc5438412a service nova] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.462906] env[62814]: DEBUG oslo_concurrency.lockutils [req-9c6cad8c-defb-4e15-a330-f187ac121fcc req-2d2716c9-1cf1-42df-8162-dfbc5438412a service nova] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.462906] env[62814]: DEBUG nova.compute.manager [req-9c6cad8c-defb-4e15-a330-f187ac121fcc req-2d2716c9-1cf1-42df-8162-dfbc5438412a service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] No waiting events found dispatching network-vif-plugged-a00402c1-ebc9-40c0-93b2-26b6fbab4a68 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 689.463141] env[62814]: WARNING nova.compute.manager [req-9c6cad8c-defb-4e15-a330-f187ac121fcc req-2d2716c9-1cf1-42df-8162-dfbc5438412a service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Received unexpected event network-vif-plugged-a00402c1-ebc9-40c0-93b2-26b6fbab4a68 for instance with vm_state building and task_state spawning. [ 689.534141] env[62814]: DEBUG nova.compute.manager [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 689.534428] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 689.535354] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d52a02-1066-482e-872c-9bba29ceb62f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.545685] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 689.545986] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8875d739-a5d6-4f47-90e0-e4f4bbf63295 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.553270] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.553270] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 689.553665] env[62814]: DEBUG nova.network.neutron [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.557019] env[62814]: DEBUG oslo_vmware.api [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 689.557019] env[62814]: value = "task-4293376" [ 689.557019] env[62814]: _type = "Task" [ 689.557019] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.569063] env[62814]: DEBUG oslo_vmware.api [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293376, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.718541] env[62814]: DEBUG nova.scheduler.client.report [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 689.726188] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 689.823582] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.823842] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.824062] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 689.824255] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.824422] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.828838] env[62814]: INFO nova.compute.manager [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Terminating instance [ 690.071594] env[62814]: DEBUG oslo_vmware.api [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293376, 'name': PowerOffVM_Task, 'duration_secs': 0.264456} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.074859] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 690.074859] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 690.074859] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c49996f-484d-4c79-bddc-998774aae059 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.124543] env[62814]: DEBUG nova.network.neutron [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.157251] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 690.157764] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 690.158120] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleting the datastore file [datastore2] a36ad785-2f33-4dbc-bc82-ab4a35020b0f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 690.158479] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-634857df-d97e-4392-9a1e-a858656e4b12 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.168231] env[62814]: DEBUG oslo_vmware.api [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 690.168231] env[62814]: value = "task-4293378" [ 690.168231] env[62814]: _type = "Task" [ 690.168231] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.179796] env[62814]: DEBUG oslo_vmware.api [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293378, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.222618] env[62814]: DEBUG oslo_concurrency.lockutils [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.186s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 690.224395] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.228s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 690.226687] env[62814]: INFO nova.compute.claims [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 690.268620] env[62814]: INFO nova.scheduler.client.report [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Deleted allocations for instance f5504590-823d-4c37-8226-26684f7957a6 [ 690.332980] env[62814]: DEBUG nova.compute.manager [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 690.333280] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 690.334759] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a0d6d9-fcd6-438d-8191-15b87d7e52d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.348021] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 690.348021] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-40942fa4-d88a-442e-96c8-d678d8241720 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.356354] env[62814]: DEBUG oslo_vmware.api [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 690.356354] env[62814]: value = "task-4293379" [ 690.356354] env[62814]: _type = "Task" [ 690.356354] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.366877] env[62814]: DEBUG oslo_vmware.api [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293379, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.396263] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.396450] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 690.396571] env[62814]: DEBUG nova.network.neutron [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 690.428209] env[62814]: DEBUG nova.network.neutron [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance_info_cache with network_info: [{"id": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "address": "fa:16:3e:93:37:3e", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa00402c1-eb", "ovs_interfaceid": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.634800] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquiring lock "9677b03f-7138-47b7-b1e8-f3714d11e550" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 690.635342] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lock "9677b03f-7138-47b7-b1e8-f3714d11e550" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 690.680699] env[62814]: DEBUG oslo_vmware.api [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293378, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.253213} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.680699] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 690.680699] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 690.680699] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.680699] env[62814]: INFO nova.compute.manager [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 690.680917] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 690.680917] env[62814]: DEBUG nova.compute.manager [-] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 690.680917] env[62814]: DEBUG nova.network.neutron [-] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.779776] env[62814]: DEBUG oslo_concurrency.lockutils [None req-82ce54a8-45ff-4f4d-a6c2-5a4f6348b21c tempest-ServerDiagnosticsTest-1981255997 tempest-ServerDiagnosticsTest-1981255997-project-member] Lock "f5504590-823d-4c37-8226-26684f7957a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.737s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 690.870939] env[62814]: DEBUG oslo_vmware.api [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293379, 'name': PowerOffVM_Task, 'duration_secs': 0.27425} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.871272] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 690.871529] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 690.871885] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e445a62c-d89e-4081-85e2-b7ccbf8c2aae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.929368] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 690.930640] env[62814]: DEBUG nova.compute.manager [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Instance network_info: |[{"id": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "address": "fa:16:3e:93:37:3e", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa00402c1-eb", "ovs_interfaceid": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 690.931886] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:37:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a00402c1-ebc9-40c0-93b2-26b6fbab4a68', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 690.948240] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Creating folder: Project (513135674ee446d19fa8c667a47138a5). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 690.948240] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9fe6c649-5c89-41ec-9c14-97c22338e1f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.949764] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 690.949956] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 690.950152] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleting the datastore file [datastore2] 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 690.950494] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78829319-fdbb-4989-881c-b0434bb1a086 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.958506] env[62814]: DEBUG oslo_vmware.api [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 690.958506] env[62814]: value = "task-4293382" [ 690.958506] env[62814]: _type = "Task" [ 690.958506] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.963711] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Created folder: Project (513135674ee446d19fa8c667a47138a5) in parent group-v845547. [ 690.963923] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Creating folder: Instances. Parent ref: group-v845606. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 690.964525] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a48a95eb-a57b-4f8e-87f3-7dcabff27972 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.970594] env[62814]: DEBUG oslo_vmware.api [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293382, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.975606] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Created folder: Instances in parent group-v845606. [ 690.976266] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 690.976447] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 690.980022] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-97999ca2-bf12-4db2-8b41-fdce036fb915 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.998276] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 690.998276] env[62814]: value = "task-4293384" [ 690.998276] env[62814]: _type = "Task" [ 690.998276] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.009567] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293384, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.236753] env[62814]: DEBUG nova.network.neutron [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance_info_cache with network_info: [{"id": "827a94b4-5864-4060-bfb5-d0e9d2281332", "address": "fa:16:3e:a9:8f:95", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.43", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap827a94b4-58", "ovs_interfaceid": "827a94b4-5864-4060-bfb5-d0e9d2281332", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.271037] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 691.271037] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35fc72ba-16c5-4efc-826b-4dd8abdf389d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.303395] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 691.303395] env[62814]: value = "task-4293385" [ 691.303395] env[62814]: _type = "Task" [ 691.303395] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.321873] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.477726] env[62814]: DEBUG oslo_vmware.api [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293382, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143324} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.478147] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 691.478364] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 691.478544] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 691.478758] env[62814]: INFO nova.compute.manager [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Took 1.15 seconds to destroy the instance on the hypervisor. [ 691.479273] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 691.483022] env[62814]: DEBUG nova.compute.manager [-] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 691.483022] env[62814]: DEBUG nova.network.neutron [-] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 691.511439] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293384, 'name': CreateVM_Task, 'duration_secs': 0.388035} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.511714] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 691.512684] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.512684] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.513049] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 691.513378] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0edbf92-06bb-47d8-b405-92445da64dbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.524215] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 691.524215] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52af5283-a316-4c07-b8b0-28bb483fd202" [ 691.524215] env[62814]: _type = "Task" [ 691.524215] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.535868] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52af5283-a316-4c07-b8b0-28bb483fd202, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.634838] env[62814]: DEBUG nova.network.neutron [-] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.742654] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.823429] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293385, 'name': PowerOffVM_Task, 'duration_secs': 0.327827} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.824226] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 691.824758] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58562fb9-f70e-4005-b651-1ba0c9aa4758 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.854969] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ceaa3d0-8784-4bf1-8cbe-0fc8e51ebd31 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.876773] env[62814]: DEBUG nova.compute.manager [req-15bcd693-9fbb-46f6-a19d-e024ace114fe req-e7aa3bfe-28b4-42a0-b031-6a109baf895b service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Received event network-changed-a00402c1-ebc9-40c0-93b2-26b6fbab4a68 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 691.876773] env[62814]: DEBUG nova.compute.manager [req-15bcd693-9fbb-46f6-a19d-e024ace114fe req-e7aa3bfe-28b4-42a0-b031-6a109baf895b service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Refreshing instance network info cache due to event network-changed-a00402c1-ebc9-40c0-93b2-26b6fbab4a68. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 691.877144] env[62814]: DEBUG oslo_concurrency.lockutils [req-15bcd693-9fbb-46f6-a19d-e024ace114fe req-e7aa3bfe-28b4-42a0-b031-6a109baf895b service nova] Acquiring lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.877246] env[62814]: DEBUG oslo_concurrency.lockutils [req-15bcd693-9fbb-46f6-a19d-e024ace114fe req-e7aa3bfe-28b4-42a0-b031-6a109baf895b service nova] Acquired lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.877443] env[62814]: DEBUG nova.network.neutron [req-15bcd693-9fbb-46f6-a19d-e024ace114fe req-e7aa3bfe-28b4-42a0-b031-6a109baf895b service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Refreshing network info cache for port a00402c1-ebc9-40c0-93b2-26b6fbab4a68 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 691.905579] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 691.905579] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-26407ed1-6d64-46c4-94d4-2bea29c55274 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.914149] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 691.914149] env[62814]: value = "task-4293386" [ 691.914149] env[62814]: _type = "Task" [ 691.914149] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.926536] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 691.927076] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 691.927178] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.927302] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.927985] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 691.928361] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c807cc34-6dfc-432f-bca3-4b8ab6695847 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.939243] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 691.939382] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 691.940124] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b46be1e-4b5e-4b72-9559-f7d943322f72 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.947798] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 691.947798] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52017e8c-5912-406c-c131-e72d3198efe4" [ 691.947798] env[62814]: _type = "Task" [ 691.947798] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.956618] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4f57a2-185f-4aa1-a608-b1ee389e3087 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.963031] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52017e8c-5912-406c-c131-e72d3198efe4, 'name': SearchDatastore_Task, 'duration_secs': 0.010763} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.964121] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-412264b9-0dc2-4ab2-b115-de1ce5616b44 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.969228] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc434917-eb39-487a-ac31-fb952964f993 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.973509] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 691.973509] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c9c281-95b4-c352-2d15-ab9f2630a0cd" [ 691.973509] env[62814]: _type = "Task" [ 691.973509] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.002259] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8455e8a-0397-450a-b33f-b5f16fe35ea7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.008326] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c9c281-95b4-c352-2d15-ab9f2630a0cd, 'name': SearchDatastore_Task, 'duration_secs': 0.010573} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.008997] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.009288] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c7cef7f3-11db-44e1-a454-98830b465b52/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. {{(pid=62814) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 692.009581] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-611185b3-e41c-41cb-94ea-deb469a829b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.015971] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f49602-38c8-4706-9082-aa74aa43e7d1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.020903] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 692.020903] env[62814]: value = "task-4293387" [ 692.020903] env[62814]: _type = "Task" [ 692.020903] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.032620] env[62814]: DEBUG nova.compute.provider_tree [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.042606] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52af5283-a316-4c07-b8b0-28bb483fd202, 'name': SearchDatastore_Task, 'duration_secs': 0.012818} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.045604] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.045843] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 692.046088] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.046234] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.046408] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 692.046727] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293387, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.047527] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a7f5592-3232-48b2-ae01-791bcb0332d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.058185] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 692.058327] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 692.059086] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04eefbec-f13f-416c-842a-938ccbee7050 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.065676] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 692.065676] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5250032e-c0c4-c77e-f559-808fa5119d0a" [ 692.065676] env[62814]: _type = "Task" [ 692.065676] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.074020] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5250032e-c0c4-c77e-f559-808fa5119d0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.139804] env[62814]: INFO nova.compute.manager [-] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Took 1.46 seconds to deallocate network for instance. [ 692.273271] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e34a16-0f5d-4e45-86aa-f5845bc7ce74 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.308021] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77b7fcc-9f70-49f0-8900-a3ba22eef991 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.314408] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance '6a592192-1b41-4be2-84a6-c3b76a4e5643' progress to 83 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 692.367105] env[62814]: DEBUG nova.network.neutron [-] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.540241] env[62814]: DEBUG nova.scheduler.client.report [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 692.543547] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293387, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.501237} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.544022] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c7cef7f3-11db-44e1-a454-98830b465b52/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. [ 692.547510] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d99a23-663b-4bb7-9a6f-6250243723c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.583553] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Reconfiguring VM instance instance-0000000c to attach disk [datastore2] c7cef7f3-11db-44e1-a454-98830b465b52/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 692.586578] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae12ac7e-a64e-46cb-ab99-ceb4c71b316a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.620901] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 692.620901] env[62814]: value = "task-4293388" [ 692.620901] env[62814]: _type = "Task" [ 692.620901] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.622027] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5250032e-c0c4-c77e-f559-808fa5119d0a, 'name': SearchDatastore_Task, 'duration_secs': 0.009999} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.625384] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f559baf7-1737-466f-a3a4-d1a0055fa068 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.637938] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 692.637938] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522cc004-4670-fe6d-f6dc-1729df62b042" [ 692.637938] env[62814]: _type = "Task" [ 692.637938] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.638370] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293388, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.648922] env[62814]: DEBUG oslo_concurrency.lockutils [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 692.649351] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522cc004-4670-fe6d-f6dc-1729df62b042, 'name': SearchDatastore_Task, 'duration_secs': 0.011374} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.649831] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.649925] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 59d21ef7-df97-49ac-9329-4c18df6dd087/59d21ef7-df97-49ac-9329-4c18df6dd087.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 692.650257] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f463a80-9315-4f94-8d26-250aeb4fda21 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.659518] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 692.659518] env[62814]: value = "task-4293389" [ 692.659518] env[62814]: _type = "Task" [ 692.659518] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.669147] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293389, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.743846] env[62814]: DEBUG nova.network.neutron [req-15bcd693-9fbb-46f6-a19d-e024ace114fe req-e7aa3bfe-28b4-42a0-b031-6a109baf895b service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updated VIF entry in instance network info cache for port a00402c1-ebc9-40c0-93b2-26b6fbab4a68. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 692.744555] env[62814]: DEBUG nova.network.neutron [req-15bcd693-9fbb-46f6-a19d-e024ace114fe req-e7aa3bfe-28b4-42a0-b031-6a109baf895b service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance_info_cache with network_info: [{"id": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "address": "fa:16:3e:93:37:3e", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa00402c1-eb", "ovs_interfaceid": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.823017] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 692.826736] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b6eff47-f77e-4e0a-8614-0d57b62dd1cf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.840824] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 692.840824] env[62814]: value = "task-4293390" [ 692.840824] env[62814]: _type = "Task" [ 692.840824] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.855717] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293390, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.869746] env[62814]: INFO nova.compute.manager [-] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Took 1.39 seconds to deallocate network for instance. [ 693.045826] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.821s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 693.048034] env[62814]: DEBUG nova.compute.manager [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 693.051341] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.224s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 693.057698] env[62814]: INFO nova.compute.claims [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.137340] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293388, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.173757] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293389, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467454} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.174105] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 59d21ef7-df97-49ac-9329-4c18df6dd087/59d21ef7-df97-49ac-9329-4c18df6dd087.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 693.174351] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 693.174583] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-170652c1-9ad9-4602-8741-3bc255e9c7b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.183658] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 693.183658] env[62814]: value = "task-4293391" [ 693.183658] env[62814]: _type = "Task" [ 693.183658] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.194963] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293391, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.248164] env[62814]: DEBUG oslo_concurrency.lockutils [req-15bcd693-9fbb-46f6-a19d-e024ace114fe req-e7aa3bfe-28b4-42a0-b031-6a109baf895b service nova] Releasing lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 693.248426] env[62814]: DEBUG nova.compute.manager [req-15bcd693-9fbb-46f6-a19d-e024ace114fe req-e7aa3bfe-28b4-42a0-b031-6a109baf895b service nova] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Received event network-vif-deleted-fb397a6f-5617-48a5-b5f6-30d9be90c5d4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 693.353417] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293390, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.379414] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 693.571201] env[62814]: DEBUG nova.compute.utils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 693.571850] env[62814]: DEBUG nova.compute.manager [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 693.574020] env[62814]: DEBUG nova.network.neutron [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 693.637470] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293388, 'name': ReconfigVM_Task, 'duration_secs': 0.568892} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.637900] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Reconfigured VM instance instance-0000000c to attach disk [datastore2] c7cef7f3-11db-44e1-a454-98830b465b52/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 693.639062] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c208c1b2-b14f-410a-ae5d-1108e21c6d0c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.644297] env[62814]: DEBUG nova.policy [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c61ebde3352e4bb18abad5ab143bbd26', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b776e7b42ebb42f09820b9005d04049a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 693.671084] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fdd5eee9-866f-4dd9-9fbc-4d743be99ca1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.697435] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293391, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075447} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.699842] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 693.699842] env[62814]: value = "task-4293392" [ 693.699842] env[62814]: _type = "Task" [ 693.699842] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.699842] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 693.699842] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890a1f42-c544-4217-b556-cfb23d32d071 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.705035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "98d134b4-b4ca-4247-a638-ad5c24a694e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 693.705249] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "98d134b4-b4ca-4247-a638-ad5c24a694e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 693.705546] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "98d134b4-b4ca-4247-a638-ad5c24a694e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 693.705634] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "98d134b4-b4ca-4247-a638-ad5c24a694e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 693.705959] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "98d134b4-b4ca-4247-a638-ad5c24a694e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 693.708371] env[62814]: INFO nova.compute.manager [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Terminating instance [ 693.743265] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] 59d21ef7-df97-49ac-9329-4c18df6dd087/59d21ef7-df97-49ac-9329-4c18df6dd087.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 693.743265] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293392, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.746818] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0763f147-d18d-4d6d-bcdf-095c968c50bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.770037] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 693.770037] env[62814]: value = "task-4293393" [ 693.770037] env[62814]: _type = "Task" [ 693.770037] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.781657] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293393, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.858032] env[62814]: DEBUG oslo_vmware.api [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293390, 'name': PowerOnVM_Task, 'duration_secs': 0.612324} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.858032] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 693.858032] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1c21146-6feb-41a2-8148-1f67f60418f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance '6a592192-1b41-4be2-84a6-c3b76a4e5643' progress to 100 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 694.076996] env[62814]: DEBUG nova.compute.manager [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 694.215644] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293392, 'name': ReconfigVM_Task, 'duration_secs': 0.187737} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.216082] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 694.216490] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3773b4aa-d6f8-40f0-b1ae-2b16cbbb155b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.228086] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 694.228086] env[62814]: value = "task-4293394" [ 694.228086] env[62814]: _type = "Task" [ 694.228086] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.241736] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293394, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.248020] env[62814]: DEBUG nova.compute.manager [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 694.248020] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.248020] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe2ed9b4-9bc0-4358-a7c7-51a054f3e4d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.256559] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 694.260440] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb8bb35d-24e1-4288-9919-32fb4ce5b9ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.283012] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293393, 'name': ReconfigVM_Task, 'duration_secs': 0.422451} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.288998] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Reconfigured VM instance instance-00000014 to attach disk [datastore2] 59d21ef7-df97-49ac-9329-4c18df6dd087/59d21ef7-df97-49ac-9329-4c18df6dd087.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 694.288998] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e9f924b-2eff-4fca-9b2a-1c751aaf10d6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.296018] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 694.296018] env[62814]: value = "task-4293396" [ 694.296018] env[62814]: _type = "Task" [ 694.296018] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.305711] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293396, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.306705] env[62814]: DEBUG nova.network.neutron [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Successfully created port: 7aee2670-a9fd-4740-9463-7492279dcfc6 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.328437] env[62814]: DEBUG nova.compute.manager [req-b4a71e43-88ff-47e4-8e04-422a78d6e485 req-8f8b0e78-7fc3-4c98-87f2-8bac146d01c7 service nova] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Received event network-vif-deleted-04ddd249-262e-42ae-95d7-a32509b50e6d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 694.350286] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 694.350514] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 694.351680] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Deleting the datastore file [datastore2] 98d134b4-b4ca-4247-a638-ad5c24a694e5 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 694.351680] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ecbc342-4f6b-4d7d-ba3e-44e2f0d03a81 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.360640] env[62814]: DEBUG oslo_vmware.api [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 694.360640] env[62814]: value = "task-4293397" [ 694.360640] env[62814]: _type = "Task" [ 694.360640] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.377868] env[62814]: DEBUG oslo_vmware.api [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293397, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.397306] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "c9e11016-e92a-459e-b5ee-b0e43ce29450" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 694.397917] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "c9e11016-e92a-459e-b5ee-b0e43ce29450" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 694.716562] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2692b5ec-5243-432a-8e1b-6128fd2095e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.724419] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329b4852-cd5b-443b-ac61-9610e21c550d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.761015] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2234ed8-b821-40f5-bdc2-788a2d2a0ed4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.765834] env[62814]: DEBUG oslo_vmware.api [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293394, 'name': PowerOnVM_Task, 'duration_secs': 0.445023} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.768014] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 694.773020] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875426e9-e7be-497f-958d-5f90cdf99a52 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.776297] env[62814]: DEBUG nova.compute.manager [None req-1ca1cf6a-09ee-4625-b71d-4ee6df2832a6 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 694.778556] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7dfe261-fff8-4f58-854d-085baa62d2aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.791152] env[62814]: DEBUG nova.compute.provider_tree [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.806681] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293396, 'name': Rename_Task, 'duration_secs': 0.168337} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.807046] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 694.807330] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8ca5417-af18-47e9-b336-e36709d7fe28 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.814448] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 694.814448] env[62814]: value = "task-4293398" [ 694.814448] env[62814]: _type = "Task" [ 694.814448] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.823780] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293398, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.876950] env[62814]: DEBUG oslo_vmware.api [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293397, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.296644} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.878221] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 694.878454] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 694.878781] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.878951] env[62814]: INFO nova.compute.manager [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Took 0.63 seconds to destroy the instance on the hypervisor. [ 694.879294] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 694.879807] env[62814]: DEBUG nova.compute.manager [-] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 694.879807] env[62814]: DEBUG nova.network.neutron [-] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 695.085956] env[62814]: DEBUG nova.compute.manager [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 695.117202] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 695.117554] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.117867] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 695.118372] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.118372] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 695.119454] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 695.119454] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 695.119454] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 695.119454] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 695.119748] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 695.119861] env[62814]: DEBUG nova.virt.hardware [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 695.121494] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731c2b54-153f-484a-89bb-067f16cb0528 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.132112] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807f9aa8-5602-43ab-a303-e6bce45e0972 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.299233] env[62814]: DEBUG nova.scheduler.client.report [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 695.332111] env[62814]: DEBUG oslo_vmware.api [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293398, 'name': PowerOnVM_Task, 'duration_secs': 0.460256} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.332111] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 695.332111] env[62814]: INFO nova.compute.manager [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Took 7.59 seconds to spawn the instance on the hypervisor. [ 695.332111] env[62814]: DEBUG nova.compute.manager [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 695.332111] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ace698c-987d-4471-99e2-d5d3dea7a8e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.678844] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquiring lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.679296] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.767827] env[62814]: DEBUG nova.network.neutron [-] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.813673] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.759s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 695.813673] env[62814]: DEBUG nova.compute.manager [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 695.817568] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.741s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.818535] env[62814]: INFO nova.compute.claims [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 695.853442] env[62814]: INFO nova.compute.manager [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Took 43.19 seconds to build instance. [ 696.260294] env[62814]: DEBUG nova.network.neutron [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Successfully updated port: 7aee2670-a9fd-4740-9463-7492279dcfc6 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 696.272529] env[62814]: INFO nova.compute.manager [-] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Took 1.39 seconds to deallocate network for instance. [ 696.324585] env[62814]: DEBUG nova.compute.utils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 696.326688] env[62814]: DEBUG nova.compute.manager [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 696.326873] env[62814]: DEBUG nova.network.neutron [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 696.357263] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7b950ee-47d7-4b16-a379-8ea64fa92e6a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.710s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 696.429785] env[62814]: DEBUG nova.policy [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '713f3b7e15404bffa24349dac34c8534', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f11d1ef620764fa4b1e2b718ac207a44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 696.651680] env[62814]: DEBUG nova.compute.manager [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Received event network-vif-deleted-05dc355c-bbf9-44a2-80dc-4a730699b8f8 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 696.651910] env[62814]: DEBUG nova.compute.manager [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Received event network-vif-plugged-7aee2670-a9fd-4740-9463-7492279dcfc6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 696.652132] env[62814]: DEBUG oslo_concurrency.lockutils [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] Acquiring lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.652339] env[62814]: DEBUG oslo_concurrency.lockutils [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] Lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 696.652502] env[62814]: DEBUG oslo_concurrency.lockutils [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] Lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 696.652685] env[62814]: DEBUG nova.compute.manager [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] No waiting events found dispatching network-vif-plugged-7aee2670-a9fd-4740-9463-7492279dcfc6 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 696.652847] env[62814]: WARNING nova.compute.manager [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Received unexpected event network-vif-plugged-7aee2670-a9fd-4740-9463-7492279dcfc6 for instance with vm_state building and task_state spawning. [ 696.653024] env[62814]: DEBUG nova.compute.manager [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Received event network-changed-7aee2670-a9fd-4740-9463-7492279dcfc6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 696.653185] env[62814]: DEBUG nova.compute.manager [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Refreshing instance network info cache due to event network-changed-7aee2670-a9fd-4740-9463-7492279dcfc6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 696.653401] env[62814]: DEBUG oslo_concurrency.lockutils [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] Acquiring lock "refresh_cache-8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.653495] env[62814]: DEBUG oslo_concurrency.lockutils [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] Acquired lock "refresh_cache-8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 696.653640] env[62814]: DEBUG nova.network.neutron [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Refreshing network info cache for port 7aee2670-a9fd-4740-9463-7492279dcfc6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 696.763406] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquiring lock "refresh_cache-8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.780743] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 696.836794] env[62814]: DEBUG nova.compute.manager [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 696.864023] env[62814]: DEBUG nova.compute.manager [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 696.956249] env[62814]: INFO nova.compute.manager [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Unrescuing [ 696.956249] env[62814]: DEBUG oslo_concurrency.lockutils [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.956249] env[62814]: DEBUG oslo_concurrency.lockutils [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 696.956249] env[62814]: DEBUG nova.network.neutron [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.099687] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "6a592192-1b41-4be2-84a6-c3b76a4e5643" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.100344] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 697.100615] env[62814]: DEBUG nova.compute.manager [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Going to confirm migration 1 {{(pid=62814) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 697.235966] env[62814]: DEBUG nova.network.neutron [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.383498] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.511708] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85eb440c-a7c3-41d4-8a44-d021a517489f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.524632] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aeee5fb-e625-463e-88da-c70e84d158b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.575772] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af5c626-f451-4faa-9c05-d494b630cb06 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.589018] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af9c542-a0cd-447a-a440-ebc8af8311a3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.611917] env[62814]: DEBUG nova.compute.provider_tree [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.616555] env[62814]: DEBUG nova.network.neutron [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Successfully created port: 2b7654ff-51f0-4a51-9b69-04d3352ddbee {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 697.646410] env[62814]: DEBUG nova.network.neutron [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.720998] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.720998] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 697.720998] env[62814]: DEBUG nova.network.neutron [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.720998] env[62814]: DEBUG nova.objects.instance [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lazy-loading 'info_cache' on Instance uuid 6a592192-1b41-4be2-84a6-c3b76a4e5643 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 697.852966] env[62814]: DEBUG nova.compute.manager [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 697.889708] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 697.889708] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 697.889708] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 697.890050] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 697.890050] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 697.890228] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 697.890567] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 697.890846] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 697.891174] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 697.891471] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 697.891778] env[62814]: DEBUG nova.virt.hardware [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 697.893474] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c02eab-b582-4004-bb53-2895aeaa9b6b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.916253] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-634b0224-b64c-4000-9f81-c1c105bfed43 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.120465] env[62814]: DEBUG nova.scheduler.client.report [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 698.129292] env[62814]: DEBUG nova.network.neutron [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Updating instance_info_cache with network_info: [{"id": "22bf5740-3f2e-45ed-be98-ac62c70cfb3b", "address": "fa:16:3e:6e:e4:ef", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap22bf5740-3f", "ovs_interfaceid": "22bf5740-3f2e-45ed-be98-ac62c70cfb3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.149496] env[62814]: DEBUG oslo_concurrency.lockutils [req-7d389e89-27b3-48d4-9233-97f80019ae70 req-b88ec463-aa2b-4981-9bf0-587fec9e68bc service nova] Releasing lock "refresh_cache-8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 698.150431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquired lock "refresh_cache-8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.150431] env[62814]: DEBUG nova.network.neutron [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 698.633113] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.817s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 698.634195] env[62814]: DEBUG nova.compute.manager [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 698.644401] env[62814]: DEBUG oslo_concurrency.lockutils [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "refresh_cache-c7cef7f3-11db-44e1-a454-98830b465b52" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 698.644401] env[62814]: DEBUG nova.objects.instance [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lazy-loading 'flavor' on Instance uuid c7cef7f3-11db-44e1-a454-98830b465b52 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 698.644401] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.133s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.645073] env[62814]: INFO nova.compute.claims [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.702122] env[62814]: DEBUG nova.network.neutron [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.719065] env[62814]: DEBUG nova.compute.manager [req-83eab98d-f9ef-4e98-a533-765258675408 req-6f1e7f8f-d782-439c-85e8-04668acf3ab6 service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Received event network-changed-a00402c1-ebc9-40c0-93b2-26b6fbab4a68 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 698.719267] env[62814]: DEBUG nova.compute.manager [req-83eab98d-f9ef-4e98-a533-765258675408 req-6f1e7f8f-d782-439c-85e8-04668acf3ab6 service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Refreshing instance network info cache due to event network-changed-a00402c1-ebc9-40c0-93b2-26b6fbab4a68. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 698.719481] env[62814]: DEBUG oslo_concurrency.lockutils [req-83eab98d-f9ef-4e98-a533-765258675408 req-6f1e7f8f-d782-439c-85e8-04668acf3ab6 service nova] Acquiring lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.719621] env[62814]: DEBUG oslo_concurrency.lockutils [req-83eab98d-f9ef-4e98-a533-765258675408 req-6f1e7f8f-d782-439c-85e8-04668acf3ab6 service nova] Acquired lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.719802] env[62814]: DEBUG nova.network.neutron [req-83eab98d-f9ef-4e98-a533-765258675408 req-6f1e7f8f-d782-439c-85e8-04668acf3ab6 service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Refreshing network info cache for port a00402c1-ebc9-40c0-93b2-26b6fbab4a68 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 698.954026] env[62814]: DEBUG nova.network.neutron [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Updating instance_info_cache with network_info: [{"id": "7aee2670-a9fd-4740-9463-7492279dcfc6", "address": "fa:16:3e:20:98:ff", "network": {"id": "a1bdb789-23d9-4db4-b159-376e397cad6c", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-687914596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b776e7b42ebb42f09820b9005d04049a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee555dfd-3d1a-4220-89cd-ffba64e4acf0", "external-id": "nsx-vlan-transportzone-88", "segmentation_id": 88, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aee2670-a9", "ovs_interfaceid": "7aee2670-a9fd-4740-9463-7492279dcfc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.133644] env[62814]: DEBUG nova.network.neutron [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance_info_cache with network_info: [{"id": "827a94b4-5864-4060-bfb5-d0e9d2281332", "address": "fa:16:3e:a9:8f:95", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.43", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap827a94b4-58", "ovs_interfaceid": "827a94b4-5864-4060-bfb5-d0e9d2281332", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.144897] env[62814]: DEBUG nova.compute.utils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 699.147607] env[62814]: DEBUG nova.compute.manager [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 699.148095] env[62814]: DEBUG nova.network.neutron [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 699.157942] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e82653-8f30-4b59-85b3-a417780eacf9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.186537] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 699.187643] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8747dc9-216a-42a2-a2f4-1bfae23e85bb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.197716] env[62814]: DEBUG oslo_vmware.api [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 699.197716] env[62814]: value = "task-4293399" [ 699.197716] env[62814]: _type = "Task" [ 699.197716] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.209137] env[62814]: DEBUG oslo_vmware.api [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293399, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.216783] env[62814]: DEBUG nova.policy [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d52f0fcc0ec42d69e6b2094145d0f28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '558682c8ea0f4887874a1763f65cb9a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 699.463022] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Releasing lock "refresh_cache-8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.463022] env[62814]: DEBUG nova.compute.manager [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Instance network_info: |[{"id": "7aee2670-a9fd-4740-9463-7492279dcfc6", "address": "fa:16:3e:20:98:ff", "network": {"id": "a1bdb789-23d9-4db4-b159-376e397cad6c", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-687914596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b776e7b42ebb42f09820b9005d04049a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee555dfd-3d1a-4220-89cd-ffba64e4acf0", "external-id": "nsx-vlan-transportzone-88", "segmentation_id": 88, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aee2670-a9", "ovs_interfaceid": "7aee2670-a9fd-4740-9463-7492279dcfc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 699.463302] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:98:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee555dfd-3d1a-4220-89cd-ffba64e4acf0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7aee2670-a9fd-4740-9463-7492279dcfc6', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 699.471036] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Creating folder: Project (b776e7b42ebb42f09820b9005d04049a). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 699.471503] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a149ee9e-9f6b-4941-ae5a-3e1073f5c1b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.490228] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Created folder: Project (b776e7b42ebb42f09820b9005d04049a) in parent group-v845547. [ 699.492815] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Creating folder: Instances. Parent ref: group-v845609. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 699.492815] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b00073d5-7ac6-490f-b083-3fe8ca26523b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.505342] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Created folder: Instances in parent group-v845609. [ 699.505342] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 699.505342] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 699.505342] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e8e30f0-e182-4e96-b218-721b154a96c2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.532349] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 699.532349] env[62814]: value = "task-4293402" [ 699.532349] env[62814]: _type = "Task" [ 699.532349] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.541054] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293402, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.637336] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "refresh_cache-6a592192-1b41-4be2-84a6-c3b76a4e5643" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.637705] env[62814]: DEBUG nova.objects.instance [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lazy-loading 'migration_context' on Instance uuid 6a592192-1b41-4be2-84a6-c3b76a4e5643 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 699.648451] env[62814]: DEBUG nova.compute.manager [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 699.662623] env[62814]: DEBUG nova.network.neutron [req-83eab98d-f9ef-4e98-a533-765258675408 req-6f1e7f8f-d782-439c-85e8-04668acf3ab6 service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updated VIF entry in instance network info cache for port a00402c1-ebc9-40c0-93b2-26b6fbab4a68. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 699.663081] env[62814]: DEBUG nova.network.neutron [req-83eab98d-f9ef-4e98-a533-765258675408 req-6f1e7f8f-d782-439c-85e8-04668acf3ab6 service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance_info_cache with network_info: [{"id": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "address": "fa:16:3e:93:37:3e", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa00402c1-eb", "ovs_interfaceid": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.712879] env[62814]: DEBUG oslo_vmware.api [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293399, 'name': PowerOffVM_Task, 'duration_secs': 0.336559} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.713900] env[62814]: DEBUG nova.network.neutron [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Successfully created port: fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 699.715702] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 699.721317] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Reconfiguring VM instance instance-0000000c to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 699.722321] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ffe697a-9f3e-4911-98e0-6ebda7b04106 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.747125] env[62814]: DEBUG oslo_vmware.api [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 699.747125] env[62814]: value = "task-4293403" [ 699.747125] env[62814]: _type = "Task" [ 699.747125] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.759829] env[62814]: DEBUG oslo_vmware.api [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293403, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.894078] env[62814]: DEBUG nova.network.neutron [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Successfully updated port: 2b7654ff-51f0-4a51-9b69-04d3352ddbee {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 700.042041] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293402, 'name': CreateVM_Task, 'duration_secs': 0.398338} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.042289] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 700.043112] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.043285] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.043653] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 700.044839] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2d7370f-3c94-40e7-9b09-1734dac9bd62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.053429] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for the task: (returnval){ [ 700.053429] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c6167c-ce52-c668-c8e4-0a4b1974c310" [ 700.053429] env[62814]: _type = "Task" [ 700.053429] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.064608] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c6167c-ce52-c668-c8e4-0a4b1974c310, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.145034] env[62814]: DEBUG nova.objects.base [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Object Instance<6a592192-1b41-4be2-84a6-c3b76a4e5643> lazy-loaded attributes: info_cache,migration_context {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 700.145034] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2def3ef-5881-4695-9ed0-e1dc71f2f271 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.170657] env[62814]: DEBUG oslo_concurrency.lockutils [req-83eab98d-f9ef-4e98-a533-765258675408 req-6f1e7f8f-d782-439c-85e8-04668acf3ab6 service nova] Releasing lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.171181] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47346450-610b-4871-968f-9a4119b95735 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.179199] env[62814]: DEBUG oslo_vmware.api [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 700.179199] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5238b24d-715c-a171-4556-3dcc4805ce93" [ 700.179199] env[62814]: _type = "Task" [ 700.179199] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.189571] env[62814]: DEBUG oslo_vmware.api [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5238b24d-715c-a171-4556-3dcc4805ce93, 'name': SearchDatastore_Task, 'duration_secs': 0.008723} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.192608] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.267240] env[62814]: DEBUG oslo_vmware.api [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293403, 'name': ReconfigVM_Task, 'duration_secs': 0.292073} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.267663] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Reconfigured VM instance instance-0000000c to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 700.267850] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 700.268126] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-994dd0ee-5c7c-4045-8220-882fd70778b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.276776] env[62814]: DEBUG oslo_vmware.api [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 700.276776] env[62814]: value = "task-4293404" [ 700.276776] env[62814]: _type = "Task" [ 700.276776] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.287395] env[62814]: DEBUG oslo_vmware.api [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293404, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.378461] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82b52c1f-ba0d-496d-a1b9-0a3ed478456d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.396022] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f988e4a-ae7d-4a5b-b494-fff0e4b669bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.401261] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "refresh_cache-4ed66b36-b6c6-4673-9c03-169a01134574" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.401433] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "refresh_cache-4ed66b36-b6c6-4673-9c03-169a01134574" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.401584] env[62814]: DEBUG nova.network.neutron [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 700.433585] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c223c470-5db0-4bfa-9eb0-d5fe632768c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.443291] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89120de-82e1-4898-aa8a-d6eb2bc207ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.458632] env[62814]: DEBUG nova.compute.provider_tree [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.565552] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c6167c-ce52-c668-c8e4-0a4b1974c310, 'name': SearchDatastore_Task, 'duration_secs': 0.012433} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.566970] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 700.566970] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 700.566970] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.566970] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.567115] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 700.567303] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb642515-f159-424c-953a-dd14aaea615d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.578543] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 700.578727] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 700.579650] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23d33c71-6473-485a-844a-7889970569fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.589035] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for the task: (returnval){ [ 700.589035] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dc86d0-fdfd-58b8-7931-27c9fa4a3a1b" [ 700.589035] env[62814]: _type = "Task" [ 700.589035] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.602433] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dc86d0-fdfd-58b8-7931-27c9fa4a3a1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.671602] env[62814]: DEBUG nova.compute.manager [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 700.695733] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 700.695733] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.695957] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 700.695986] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.696131] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 700.696422] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 700.696503] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 700.696756] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 700.696830] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 700.696985] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 700.697199] env[62814]: DEBUG nova.virt.hardware [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 700.698688] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bcc6ff-daed-4d97-b098-37c87da938a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.708632] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3785d548-7cfc-46f8-88f9-7da38197659d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.794891] env[62814]: DEBUG oslo_vmware.api [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293404, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.900581] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquiring lock "d366a755-49b4-427b-8564-d8572a7fbbb7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.900789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lock "d366a755-49b4-427b-8564-d8572a7fbbb7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 700.902565] env[62814]: DEBUG nova.compute.manager [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Received event network-vif-plugged-2b7654ff-51f0-4a51-9b69-04d3352ddbee {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 700.903148] env[62814]: DEBUG oslo_concurrency.lockutils [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] Acquiring lock "4ed66b36-b6c6-4673-9c03-169a01134574-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.903148] env[62814]: DEBUG oslo_concurrency.lockutils [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] Lock "4ed66b36-b6c6-4673-9c03-169a01134574-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 700.903148] env[62814]: DEBUG oslo_concurrency.lockutils [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] Lock "4ed66b36-b6c6-4673-9c03-169a01134574-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 700.903317] env[62814]: DEBUG nova.compute.manager [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] No waiting events found dispatching network-vif-plugged-2b7654ff-51f0-4a51-9b69-04d3352ddbee {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 700.903420] env[62814]: WARNING nova.compute.manager [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Received unexpected event network-vif-plugged-2b7654ff-51f0-4a51-9b69-04d3352ddbee for instance with vm_state building and task_state spawning. [ 700.903580] env[62814]: DEBUG nova.compute.manager [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Received event network-changed-2b7654ff-51f0-4a51-9b69-04d3352ddbee {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 700.903716] env[62814]: DEBUG nova.compute.manager [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Refreshing instance network info cache due to event network-changed-2b7654ff-51f0-4a51-9b69-04d3352ddbee. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 700.904129] env[62814]: DEBUG oslo_concurrency.lockutils [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] Acquiring lock "refresh_cache-4ed66b36-b6c6-4673-9c03-169a01134574" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.964159] env[62814]: DEBUG nova.scheduler.client.report [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 700.974587] env[62814]: DEBUG nova.network.neutron [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.101301] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dc86d0-fdfd-58b8-7931-27c9fa4a3a1b, 'name': SearchDatastore_Task, 'duration_secs': 0.0127} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.102113] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e84643be-ef33-411a-b903-2f74f4831e20 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.108867] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for the task: (returnval){ [ 701.108867] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5296946d-80fe-03d5-5bb5-5ac53e7328b8" [ 701.108867] env[62814]: _type = "Task" [ 701.108867] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.119177] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5296946d-80fe-03d5-5bb5-5ac53e7328b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.159679] env[62814]: DEBUG nova.network.neutron [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Updating instance_info_cache with network_info: [{"id": "2b7654ff-51f0-4a51-9b69-04d3352ddbee", "address": "fa:16:3e:2c:cb:e3", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b7654ff-51", "ovs_interfaceid": "2b7654ff-51f0-4a51-9b69-04d3352ddbee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.187657] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "127fee64-fd56-4a23-bdd2-18c817898fd5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 701.188028] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "127fee64-fd56-4a23-bdd2-18c817898fd5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 701.290797] env[62814]: DEBUG oslo_vmware.api [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293404, 'name': PowerOnVM_Task, 'duration_secs': 0.539182} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.290797] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 701.290797] env[62814]: DEBUG nova.compute.manager [None req-529a8ce7-2ebf-400d-9665-86824525e9e4 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 701.291848] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8558f8-51c5-4334-b5d3-48de2bbb6988 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.435151] env[62814]: DEBUG nova.compute.manager [req-e886a787-1e8c-4d80-aa34-1114d756b893 req-dcee0519-94b7-443e-a9bd-64f45ad59717 service nova] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Received event network-vif-plugged-fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 701.435458] env[62814]: DEBUG oslo_concurrency.lockutils [req-e886a787-1e8c-4d80-aa34-1114d756b893 req-dcee0519-94b7-443e-a9bd-64f45ad59717 service nova] Acquiring lock "29e69c7c-08f7-4da4-9509-02a94bf971bc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 701.435761] env[62814]: DEBUG oslo_concurrency.lockutils [req-e886a787-1e8c-4d80-aa34-1114d756b893 req-dcee0519-94b7-443e-a9bd-64f45ad59717 service nova] Lock "29e69c7c-08f7-4da4-9509-02a94bf971bc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 701.435937] env[62814]: DEBUG oslo_concurrency.lockutils [req-e886a787-1e8c-4d80-aa34-1114d756b893 req-dcee0519-94b7-443e-a9bd-64f45ad59717 service nova] Lock "29e69c7c-08f7-4da4-9509-02a94bf971bc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 701.436156] env[62814]: DEBUG nova.compute.manager [req-e886a787-1e8c-4d80-aa34-1114d756b893 req-dcee0519-94b7-443e-a9bd-64f45ad59717 service nova] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] No waiting events found dispatching network-vif-plugged-fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 701.436754] env[62814]: WARNING nova.compute.manager [req-e886a787-1e8c-4d80-aa34-1114d756b893 req-dcee0519-94b7-443e-a9bd-64f45ad59717 service nova] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Received unexpected event network-vif-plugged-fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4 for instance with vm_state building and task_state spawning. [ 701.468685] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.825s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 701.469266] env[62814]: DEBUG nova.compute.manager [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 701.472489] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.499s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 701.474408] env[62814]: INFO nova.compute.claims [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.546405] env[62814]: DEBUG nova.network.neutron [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Successfully updated port: fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 701.620541] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5296946d-80fe-03d5-5bb5-5ac53e7328b8, 'name': SearchDatastore_Task, 'duration_secs': 0.014216} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.620541] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 701.620740] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4/8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 701.621000] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b150e3b3-dbde-4945-99fc-67192af9ee75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.628791] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for the task: (returnval){ [ 701.628791] env[62814]: value = "task-4293405" [ 701.628791] env[62814]: _type = "Task" [ 701.628791] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.637565] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293405, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.662323] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "refresh_cache-4ed66b36-b6c6-4673-9c03-169a01134574" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 701.662670] env[62814]: DEBUG nova.compute.manager [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Instance network_info: |[{"id": "2b7654ff-51f0-4a51-9b69-04d3352ddbee", "address": "fa:16:3e:2c:cb:e3", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b7654ff-51", "ovs_interfaceid": "2b7654ff-51f0-4a51-9b69-04d3352ddbee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 701.663037] env[62814]: DEBUG oslo_concurrency.lockutils [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] Acquired lock "refresh_cache-4ed66b36-b6c6-4673-9c03-169a01134574" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 701.663243] env[62814]: DEBUG nova.network.neutron [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Refreshing network info cache for port 2b7654ff-51f0-4a51-9b69-04d3352ddbee {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 701.664545] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:cb:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b7654ff-51f0-4a51-9b69-04d3352ddbee', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 701.676269] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Creating folder: Project (f11d1ef620764fa4b1e2b718ac207a44). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 701.677412] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03c168af-cecd-442a-bac3-c6161361fd31 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.693314] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Created folder: Project (f11d1ef620764fa4b1e2b718ac207a44) in parent group-v845547. [ 701.693572] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Creating folder: Instances. Parent ref: group-v845612. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 701.693758] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e30e698e-77cd-473b-b2d6-55361f9e18d8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.703535] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Created folder: Instances in parent group-v845612. [ 701.703736] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 701.703929] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 701.704421] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-48a2c1de-8c8b-4015-be56-f4e107db4d16 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.723469] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 701.723469] env[62814]: value = "task-4293408" [ 701.723469] env[62814]: _type = "Task" [ 701.723469] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.731379] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293408, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.977188] env[62814]: DEBUG nova.compute.utils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 701.977782] env[62814]: DEBUG nova.compute.manager [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 701.980467] env[62814]: DEBUG nova.network.neutron [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 702.049417] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "refresh_cache-29e69c7c-08f7-4da4-9509-02a94bf971bc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.049417] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired lock "refresh_cache-29e69c7c-08f7-4da4-9509-02a94bf971bc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.049645] env[62814]: DEBUG nova.network.neutron [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.081130] env[62814]: DEBUG nova.policy [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '713f3b7e15404bffa24349dac34c8534', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f11d1ef620764fa4b1e2b718ac207a44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 702.145249] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293405, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.239371] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293408, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.461543] env[62814]: DEBUG nova.network.neutron [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Updated VIF entry in instance network info cache for port 2b7654ff-51f0-4a51-9b69-04d3352ddbee. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 702.462229] env[62814]: DEBUG nova.network.neutron [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Updating instance_info_cache with network_info: [{"id": "2b7654ff-51f0-4a51-9b69-04d3352ddbee", "address": "fa:16:3e:2c:cb:e3", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b7654ff-51", "ovs_interfaceid": "2b7654ff-51f0-4a51-9b69-04d3352ddbee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.486139] env[62814]: DEBUG nova.compute.manager [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 702.513245] env[62814]: DEBUG nova.network.neutron [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Successfully created port: 44e24b3d-908f-4ded-8f46-262fb433c4a6 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.548027] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 702.549641] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 702.598949] env[62814]: DEBUG nova.network.neutron [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.645061] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293405, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.69631} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.645338] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4/8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 702.645546] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 702.645794] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a72a740-b4c3-4117-a568-392578848a0e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.657049] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for the task: (returnval){ [ 702.657049] env[62814]: value = "task-4293409" [ 702.657049] env[62814]: _type = "Task" [ 702.657049] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.665217] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293409, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.738113] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293408, 'name': CreateVM_Task, 'duration_secs': 0.806223} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.738324] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 702.739254] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.739254] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.740167] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 702.740167] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f19380f-a687-4b12-b239-23884bc8767f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.746892] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 702.746892] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cf0bf8-a47c-1ec3-e06e-2d40392f57d4" [ 702.746892] env[62814]: _type = "Task" [ 702.746892] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.756945] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cf0bf8-a47c-1ec3-e06e-2d40392f57d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.765249] env[62814]: DEBUG nova.network.neutron [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Updating instance_info_cache with network_info: [{"id": "fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4", "address": "fa:16:3e:03:1f:bc", "network": {"id": "88bc6d73-6bc2-47d0-8e78-633cbe326db2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-765790142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558682c8ea0f4887874a1763f65cb9a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb9b36b9-9a", "ovs_interfaceid": "fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.971563] env[62814]: DEBUG oslo_concurrency.lockutils [req-82082a64-afbb-4c46-b369-3d264453b3fb req-ec653009-9f02-44cb-9667-07b8a5e1e737 service nova] Releasing lock "refresh_cache-4ed66b36-b6c6-4673-9c03-169a01134574" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.061160] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 703.061426] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 703.061628] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 703.061820] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 703.062066] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 703.062242] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 703.062385] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 703.062527] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 703.128324] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34db1277-3209-494e-89d2-fb6ca5a89960 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.136353] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-799d9085-0bfc-4411-947c-65f51b326976 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.172541] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2898f8-fb09-4ba5-b0c8-9a7a139d2b71 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.180525] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293409, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075287} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.182719] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 703.183610] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322dd071-a4b3-49cc-8cad-af4910e17bd5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.187184] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136990c8-8c03-4a82-9f5e-c7a68ca611ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.201692] env[62814]: DEBUG nova.compute.provider_tree [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.222632] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Reconfiguring VM instance instance-00000015 to attach disk [datastore2] 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4/8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 703.223880] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c67743ca-00b8-4d97-92c7-161a3e0d2462 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.252603] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for the task: (returnval){ [ 703.252603] env[62814]: value = "task-4293410" [ 703.252603] env[62814]: _type = "Task" [ 703.252603] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.260009] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cf0bf8-a47c-1ec3-e06e-2d40392f57d4, 'name': SearchDatastore_Task, 'duration_secs': 0.01519} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.260663] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.260875] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 703.261130] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.261275] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.261450] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 703.261995] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf066473-08cf-4980-a2ac-6c657997a1e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.266793] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293410, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.270747] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Releasing lock "refresh_cache-29e69c7c-08f7-4da4-9509-02a94bf971bc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.271229] env[62814]: DEBUG nova.compute.manager [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Instance network_info: |[{"id": "fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4", "address": "fa:16:3e:03:1f:bc", "network": {"id": "88bc6d73-6bc2-47d0-8e78-633cbe326db2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-765790142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558682c8ea0f4887874a1763f65cb9a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb9b36b9-9a", "ovs_interfaceid": "fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 703.271685] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:1f:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '209639b9-c313-4b35-86dc-dccd744d174a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 703.279488] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Creating folder: Project (558682c8ea0f4887874a1763f65cb9a5). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 703.281307] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ec7f6f44-432b-49ac-8825-9b7c8b4bf776 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.283128] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 703.283346] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 703.284089] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8840b61f-00f6-451b-9a48-17340ddd853c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.290126] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 703.290126] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52217fbf-b454-2e42-26d1-b230715e6cc5" [ 703.290126] env[62814]: _type = "Task" [ 703.290126] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.297867] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52217fbf-b454-2e42-26d1-b230715e6cc5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.300136] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Created folder: Project (558682c8ea0f4887874a1763f65cb9a5) in parent group-v845547. [ 703.300322] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Creating folder: Instances. Parent ref: group-v845615. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 703.300557] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e1fa33e0-d64f-4ac1-a960-01f6f3ab9728 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.311125] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Created folder: Instances in parent group-v845615. [ 703.311315] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 703.311533] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 703.311757] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e9a8f2e-44bc-4f66-9b9b-088276a2d47e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.330443] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 703.330443] env[62814]: value = "task-4293413" [ 703.330443] env[62814]: _type = "Task" [ 703.330443] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.338531] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293413, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.465332] env[62814]: DEBUG nova.compute.manager [req-07296dae-cd73-4e6b-ad60-1cfca6a7d391 req-90d828ca-90e4-4e20-8703-5caf7c126712 service nova] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Received event network-changed-fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 703.465537] env[62814]: DEBUG nova.compute.manager [req-07296dae-cd73-4e6b-ad60-1cfca6a7d391 req-90d828ca-90e4-4e20-8703-5caf7c126712 service nova] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Refreshing instance network info cache due to event network-changed-fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 703.465729] env[62814]: DEBUG oslo_concurrency.lockutils [req-07296dae-cd73-4e6b-ad60-1cfca6a7d391 req-90d828ca-90e4-4e20-8703-5caf7c126712 service nova] Acquiring lock "refresh_cache-29e69c7c-08f7-4da4-9509-02a94bf971bc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.466014] env[62814]: DEBUG oslo_concurrency.lockutils [req-07296dae-cd73-4e6b-ad60-1cfca6a7d391 req-90d828ca-90e4-4e20-8703-5caf7c126712 service nova] Acquired lock "refresh_cache-29e69c7c-08f7-4da4-9509-02a94bf971bc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.466310] env[62814]: DEBUG nova.network.neutron [req-07296dae-cd73-4e6b-ad60-1cfca6a7d391 req-90d828ca-90e4-4e20-8703-5caf7c126712 service nova] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Refreshing network info cache for port fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 703.498612] env[62814]: DEBUG nova.compute.manager [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 703.527482] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 703.527770] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.527928] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 703.528133] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.528278] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 703.528421] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 703.528657] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 703.528891] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 703.529041] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 703.529211] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 703.529412] env[62814]: DEBUG nova.virt.hardware [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 703.530291] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f1c57e-35aa-4a10-a6fc-397b6b96c126 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.538603] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae5e17d-f112-43ee-8d1a-f2eeaa733117 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.568928] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 703.725966] env[62814]: DEBUG nova.scheduler.client.report [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 703.764088] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293410, 'name': ReconfigVM_Task, 'duration_secs': 0.428623} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.764088] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Reconfigured VM instance instance-00000015 to attach disk [datastore2] 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4/8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 703.764564] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f620a14-9b90-4c89-a987-3e67f5c76d00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.771233] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for the task: (returnval){ [ 703.771233] env[62814]: value = "task-4293414" [ 703.771233] env[62814]: _type = "Task" [ 703.771233] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.779659] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293414, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.801352] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52217fbf-b454-2e42-26d1-b230715e6cc5, 'name': SearchDatastore_Task, 'duration_secs': 0.023158} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.802472] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2f52846-ae4b-4c2b-9b82-ef6dc388ea83 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.809363] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 703.809363] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d223ed-1e8f-61be-c2cc-7ea3cc94344b" [ 703.809363] env[62814]: _type = "Task" [ 703.809363] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.818695] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d223ed-1e8f-61be-c2cc-7ea3cc94344b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.841922] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293413, 'name': CreateVM_Task, 'duration_secs': 0.369778} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.842311] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 703.843132] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.843240] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 703.843569] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 703.843845] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b4fa7e9-5991-492d-9d82-7de743303922 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.849171] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 703.849171] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d4e9a6-bc56-d4e0-4fad-f88f2d6cb531" [ 703.849171] env[62814]: _type = "Task" [ 703.849171] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.857213] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d4e9a6-bc56-d4e0-4fad-f88f2d6cb531, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.233027] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.759s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 704.233027] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 704.235720] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.844s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 704.236557] env[62814]: DEBUG nova.objects.instance [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lazy-loading 'resources' on Instance uuid 296f6c11-7108-42e6-8ada-5d8c08b00da6 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 704.295907] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293414, 'name': Rename_Task, 'duration_secs': 0.160836} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.296473] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 704.297070] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1fadd67-02c4-42ab-921f-554c696b2849 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.304740] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for the task: (returnval){ [ 704.304740] env[62814]: value = "task-4293415" [ 704.304740] env[62814]: _type = "Task" [ 704.304740] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.318977] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293415, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.326631] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d223ed-1e8f-61be-c2cc-7ea3cc94344b, 'name': SearchDatastore_Task, 'duration_secs': 0.010527} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.327079] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.328319] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 704.328319] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-259eee5a-524c-436c-8b81-56d1e4c9f451 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.333784] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 704.333784] env[62814]: value = "task-4293416" [ 704.333784] env[62814]: _type = "Task" [ 704.333784] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.342538] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293416, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.361797] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d4e9a6-bc56-d4e0-4fad-f88f2d6cb531, 'name': SearchDatastore_Task, 'duration_secs': 0.010673} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.362805] env[62814]: DEBUG nova.network.neutron [req-07296dae-cd73-4e6b-ad60-1cfca6a7d391 req-90d828ca-90e4-4e20-8703-5caf7c126712 service nova] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Updated VIF entry in instance network info cache for port fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 704.363164] env[62814]: DEBUG nova.network.neutron [req-07296dae-cd73-4e6b-ad60-1cfca6a7d391 req-90d828ca-90e4-4e20-8703-5caf7c126712 service nova] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Updating instance_info_cache with network_info: [{"id": "fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4", "address": "fa:16:3e:03:1f:bc", "network": {"id": "88bc6d73-6bc2-47d0-8e78-633cbe326db2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-765790142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558682c8ea0f4887874a1763f65cb9a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb9b36b9-9a", "ovs_interfaceid": "fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.364995] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.365422] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 704.366505] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.366734] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 704.367021] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 704.368167] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b8988eb-b7c5-463f-ad4a-741ee17fe04c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.386070] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 704.386279] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 704.387075] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9cab5a0-e112-42a5-86c9-553fbbb7f4cc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.393098] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 704.393098] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e901b4-2c6d-9942-67bd-1508052eadea" [ 704.393098] env[62814]: _type = "Task" [ 704.393098] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.403011] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e901b4-2c6d-9942-67bd-1508052eadea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.453258] env[62814]: DEBUG nova.network.neutron [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Successfully updated port: 44e24b3d-908f-4ded-8f46-262fb433c4a6 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 704.740554] env[62814]: DEBUG nova.compute.utils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 704.744872] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 704.745164] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 704.821325] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293415, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.846360] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293416, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.863751] env[62814]: DEBUG nova.policy [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a2e083ae164419fa78addf5da979950', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3687fbff426473b9d6135208426b66e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 704.870911] env[62814]: DEBUG oslo_concurrency.lockutils [req-07296dae-cd73-4e6b-ad60-1cfca6a7d391 req-90d828ca-90e4-4e20-8703-5caf7c126712 service nova] Releasing lock "refresh_cache-29e69c7c-08f7-4da4-9509-02a94bf971bc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.915012] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e901b4-2c6d-9942-67bd-1508052eadea, 'name': SearchDatastore_Task, 'duration_secs': 0.010226} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.915871] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f4e71ac-143e-45a9-a753-816bc92e08b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.921964] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 704.921964] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f43e3a-05ef-40d7-e655-99fce82ec5df" [ 704.921964] env[62814]: _type = "Task" [ 704.921964] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.937060] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f43e3a-05ef-40d7-e655-99fce82ec5df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.957187] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "refresh_cache-f5ad9d70-75fb-4881-8853-5ede4d0903f2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.957394] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "refresh_cache-f5ad9d70-75fb-4881-8853-5ede4d0903f2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 704.957550] env[62814]: DEBUG nova.network.neutron [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.252683] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 705.316950] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293415, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.322115] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Successfully created port: 6be6bb54-d3f3-4068-9322-ccdb97c8470d {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 705.344660] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293416, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553608} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.347185] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 705.347404] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 705.347821] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6bfc92cf-b9b8-48ad-8e02-617711f8bd10 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.355260] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 705.355260] env[62814]: value = "task-4293417" [ 705.355260] env[62814]: _type = "Task" [ 705.355260] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.367908] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293417, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.416508] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0831e6-7bb1-4ef5-b9f0-7afab9a995a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.427342] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c07d05d-bd05-415c-b2b3-80da06e085f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.435848] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f43e3a-05ef-40d7-e655-99fce82ec5df, 'name': SearchDatastore_Task, 'duration_secs': 0.01492} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.460613] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 705.460932] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 29e69c7c-08f7-4da4-9509-02a94bf971bc/29e69c7c-08f7-4da4-9509-02a94bf971bc.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 705.463512] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7cf5145d-f137-44ba-aaeb-f9c605a2bfd2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.465853] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58246f42-4e37-4a08-b555-8042a5ceb0fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.476475] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b2a6c4-7194-4cc3-9a74-fa171209239b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.480257] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 705.480257] env[62814]: value = "task-4293418" [ 705.480257] env[62814]: _type = "Task" [ 705.480257] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.494826] env[62814]: DEBUG nova.compute.provider_tree [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.501888] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293418, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.520495] env[62814]: DEBUG nova.network.neutron [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.661808] env[62814]: DEBUG nova.compute.manager [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Received event network-vif-plugged-44e24b3d-908f-4ded-8f46-262fb433c4a6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 705.662094] env[62814]: DEBUG oslo_concurrency.lockutils [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] Acquiring lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.662351] env[62814]: DEBUG oslo_concurrency.lockutils [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] Lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.662518] env[62814]: DEBUG oslo_concurrency.lockutils [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] Lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 705.662681] env[62814]: DEBUG nova.compute.manager [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] No waiting events found dispatching network-vif-plugged-44e24b3d-908f-4ded-8f46-262fb433c4a6 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 705.662872] env[62814]: WARNING nova.compute.manager [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Received unexpected event network-vif-plugged-44e24b3d-908f-4ded-8f46-262fb433c4a6 for instance with vm_state building and task_state spawning. [ 705.664984] env[62814]: DEBUG nova.compute.manager [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Received event network-changed-44e24b3d-908f-4ded-8f46-262fb433c4a6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 705.664984] env[62814]: DEBUG nova.compute.manager [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Refreshing instance network info cache due to event network-changed-44e24b3d-908f-4ded-8f46-262fb433c4a6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 705.664984] env[62814]: DEBUG oslo_concurrency.lockutils [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] Acquiring lock "refresh_cache-f5ad9d70-75fb-4881-8853-5ede4d0903f2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.725041] env[62814]: DEBUG nova.network.neutron [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Updating instance_info_cache with network_info: [{"id": "44e24b3d-908f-4ded-8f46-262fb433c4a6", "address": "fa:16:3e:d4:8a:d1", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44e24b3d-90", "ovs_interfaceid": "44e24b3d-908f-4ded-8f46-262fb433c4a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.819102] env[62814]: DEBUG oslo_vmware.api [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293415, 'name': PowerOnVM_Task, 'duration_secs': 1.310039} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.819102] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 705.819315] env[62814]: INFO nova.compute.manager [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Took 10.73 seconds to spawn the instance on the hypervisor. [ 705.819467] env[62814]: DEBUG nova.compute.manager [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 705.820427] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56bf112-833d-46f1-82c3-847dcf1d967d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.866238] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293417, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078248} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.866336] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 705.867186] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3763852-e8dd-4b5d-b4f3-429f3c5d366f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.889516] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 705.889885] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-649520b9-29ca-482c-9ae7-ce99c57b62bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.911442] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 705.911442] env[62814]: value = "task-4293419" [ 705.911442] env[62814]: _type = "Task" [ 705.911442] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.920270] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293419, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.990635] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293418, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.001141] env[62814]: DEBUG nova.scheduler.client.report [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 706.230040] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "refresh_cache-f5ad9d70-75fb-4881-8853-5ede4d0903f2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 706.230040] env[62814]: DEBUG nova.compute.manager [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Instance network_info: |[{"id": "44e24b3d-908f-4ded-8f46-262fb433c4a6", "address": "fa:16:3e:d4:8a:d1", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44e24b3d-90", "ovs_interfaceid": "44e24b3d-908f-4ded-8f46-262fb433c4a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 706.230207] env[62814]: DEBUG oslo_concurrency.lockutils [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] Acquired lock "refresh_cache-f5ad9d70-75fb-4881-8853-5ede4d0903f2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 706.230207] env[62814]: DEBUG nova.network.neutron [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Refreshing network info cache for port 44e24b3d-908f-4ded-8f46-262fb433c4a6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 706.230207] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:8a:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44e24b3d-908f-4ded-8f46-262fb433c4a6', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 706.238181] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 706.239568] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 706.239958] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-928b585b-59c0-4b05-b25e-2ab27082fdd4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.260810] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.261093] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.262999] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 706.267607] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 706.267607] env[62814]: value = "task-4293420" [ 706.267607] env[62814]: _type = "Task" [ 706.267607] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.275685] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293420, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.292485] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 706.292735] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 706.293451] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 706.293451] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 706.293451] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 706.293451] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 706.293617] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 706.293708] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 706.293882] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 706.294071] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 706.294270] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 706.295315] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ed0602-f940-46bf-be0f-26bba343ff00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.303139] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffe8ee5-ee92-4c7c-b461-04f6dc3aa546 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.341212] env[62814]: INFO nova.compute.manager [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Took 50.37 seconds to build instance. [ 706.423756] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293419, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.492427] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293418, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.661699} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.494113] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 29e69c7c-08f7-4da4-9509-02a94bf971bc/29e69c7c-08f7-4da4-9509-02a94bf971bc.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 706.494113] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 706.494113] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1b77726-b37c-4dd9-ad91-abbd5ed0fb28 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.500265] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 706.500265] env[62814]: value = "task-4293421" [ 706.500265] env[62814]: _type = "Task" [ 706.500265] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.503959] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.268s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.506640] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.468s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.508111] env[62814]: INFO nova.compute.claims [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.519026] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293421, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.532691] env[62814]: INFO nova.scheduler.client.report [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Deleted allocations for instance 296f6c11-7108-42e6-8ada-5d8c08b00da6 [ 706.776578] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293420, 'name': CreateVM_Task, 'duration_secs': 0.361359} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.776784] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 706.777534] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.777722] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 706.778074] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 706.778344] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78471669-8538-4b97-b655-a0cf5f3ee12b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.783850] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 706.783850] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d126da-a7b6-85f1-c351-2fdc8f51fcab" [ 706.783850] env[62814]: _type = "Task" [ 706.783850] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.792204] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d126da-a7b6-85f1-c351-2fdc8f51fcab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.843433] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6127003c-3d11-4d27-b0fa-486046c7ea88 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.884s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.922416] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293419, 'name': ReconfigVM_Task, 'duration_secs': 0.973931} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.922742] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Reconfigured VM instance instance-00000016 to attach disk [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 706.923394] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad82fe0c-0eee-4ec8-862f-65c088241dca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.929806] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 706.929806] env[62814]: value = "task-4293422" [ 706.929806] env[62814]: _type = "Task" [ 706.929806] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.943560] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293422, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.010162] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293421, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067379} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.010447] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 707.011279] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544eac43-866f-43ca-b04c-914beb46c956 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.035894] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] 29e69c7c-08f7-4da4-9509-02a94bf971bc/29e69c7c-08f7-4da4-9509-02a94bf971bc.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 707.036893] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Successfully updated port: 6be6bb54-d3f3-4068-9322-ccdb97c8470d {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 707.038556] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-326242ac-8547-40bc-b1a2-39183c1ad16a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.059764] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "refresh_cache-e6af4651-9f3a-4ce0-add8-06f1cfef255f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.060123] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired lock "refresh_cache-e6af4651-9f3a-4ce0-add8-06f1cfef255f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 707.060123] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.061895] env[62814]: DEBUG nova.compute.manager [req-a1bab209-d973-4e8a-9a9f-98bbd8957c24 req-95bb2660-8631-425d-ad49-50c9e07c99b7 service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Received event network-changed-7aee2670-a9fd-4740-9463-7492279dcfc6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 707.062078] env[62814]: DEBUG nova.compute.manager [req-a1bab209-d973-4e8a-9a9f-98bbd8957c24 req-95bb2660-8631-425d-ad49-50c9e07c99b7 service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Refreshing instance network info cache due to event network-changed-7aee2670-a9fd-4740-9463-7492279dcfc6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 707.062288] env[62814]: DEBUG oslo_concurrency.lockutils [req-a1bab209-d973-4e8a-9a9f-98bbd8957c24 req-95bb2660-8631-425d-ad49-50c9e07c99b7 service nova] Acquiring lock "refresh_cache-8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.062391] env[62814]: DEBUG oslo_concurrency.lockutils [req-a1bab209-d973-4e8a-9a9f-98bbd8957c24 req-95bb2660-8631-425d-ad49-50c9e07c99b7 service nova] Acquired lock "refresh_cache-8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 707.062539] env[62814]: DEBUG nova.network.neutron [req-a1bab209-d973-4e8a-9a9f-98bbd8957c24 req-95bb2660-8631-425d-ad49-50c9e07c99b7 service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Refreshing network info cache for port 7aee2670-a9fd-4740-9463-7492279dcfc6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 707.064115] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c121ea68-5112-447e-8f0d-0db96593a6ba tempest-ServersAaction247Test-1068619639 tempest-ServersAaction247Test-1068619639-project-member] Lock "296f6c11-7108-42e6-8ada-5d8c08b00da6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.326s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 707.070648] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 707.070648] env[62814]: value = "task-4293423" [ 707.070648] env[62814]: _type = "Task" [ 707.070648] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.079548] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293423, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.138078] env[62814]: DEBUG nova.network.neutron [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Updated VIF entry in instance network info cache for port 44e24b3d-908f-4ded-8f46-262fb433c4a6. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 707.138078] env[62814]: DEBUG nova.network.neutron [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Updating instance_info_cache with network_info: [{"id": "44e24b3d-908f-4ded-8f46-262fb433c4a6", "address": "fa:16:3e:d4:8a:d1", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44e24b3d-90", "ovs_interfaceid": "44e24b3d-908f-4ded-8f46-262fb433c4a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.294603] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d126da-a7b6-85f1-c351-2fdc8f51fcab, 'name': SearchDatastore_Task, 'duration_secs': 0.016197} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.294980] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 707.295182] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 707.295420] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.295565] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 707.295742] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 707.296011] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00efc268-6d61-434a-ab72-1a9816817ce9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.309492] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 707.309704] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 707.310477] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c148d3e-36f6-4dd0-ab0d-2b1250bf5208 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.316006] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 707.316006] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5269f00b-1102-f75d-0cc8-af1319f69d29" [ 707.316006] env[62814]: _type = "Task" [ 707.316006] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.324255] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5269f00b-1102-f75d-0cc8-af1319f69d29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.346400] env[62814]: DEBUG nova.compute.manager [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 707.440732] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293422, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.583586] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293423, 'name': ReconfigVM_Task, 'duration_secs': 0.308756} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.584912] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Reconfigured VM instance instance-00000017 to attach disk [datastore2] 29e69c7c-08f7-4da4-9509-02a94bf971bc/29e69c7c-08f7-4da4-9509-02a94bf971bc.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 707.587172] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c56f028c-c5ab-4123-abb2-098eae23a59a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.597018] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 707.597018] env[62814]: value = "task-4293424" [ 707.597018] env[62814]: _type = "Task" [ 707.597018] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.605134] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.612412] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293424, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.640135] env[62814]: DEBUG oslo_concurrency.lockutils [req-55c2d328-7468-4c26-930a-6aac07adfbd6 req-32a92b87-ba3c-4b98-b188-2c40a2b60c61 service nova] Releasing lock "refresh_cache-f5ad9d70-75fb-4881-8853-5ede4d0903f2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 707.757176] env[62814]: DEBUG nova.compute.manager [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Received event network-vif-plugged-6be6bb54-d3f3-4068-9322-ccdb97c8470d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 707.757176] env[62814]: DEBUG oslo_concurrency.lockutils [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] Acquiring lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 707.757382] env[62814]: DEBUG oslo_concurrency.lockutils [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] Lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 707.757537] env[62814]: DEBUG oslo_concurrency.lockutils [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] Lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 707.757690] env[62814]: DEBUG nova.compute.manager [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] No waiting events found dispatching network-vif-plugged-6be6bb54-d3f3-4068-9322-ccdb97c8470d {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 707.757850] env[62814]: WARNING nova.compute.manager [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Received unexpected event network-vif-plugged-6be6bb54-d3f3-4068-9322-ccdb97c8470d for instance with vm_state building and task_state spawning. [ 707.759213] env[62814]: DEBUG nova.compute.manager [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Received event network-changed-6be6bb54-d3f3-4068-9322-ccdb97c8470d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 707.759635] env[62814]: DEBUG nova.compute.manager [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Refreshing instance network info cache due to event network-changed-6be6bb54-d3f3-4068-9322-ccdb97c8470d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 707.759635] env[62814]: DEBUG oslo_concurrency.lockutils [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] Acquiring lock "refresh_cache-e6af4651-9f3a-4ce0-add8-06f1cfef255f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.806983] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Updating instance_info_cache with network_info: [{"id": "6be6bb54-d3f3-4068-9322-ccdb97c8470d", "address": "fa:16:3e:a2:b6:05", "network": {"id": "af150ca9-547c-4640-ae29-015a75200623", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-81489590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3687fbff426473b9d6135208426b66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6be6bb54-d3", "ovs_interfaceid": "6be6bb54-d3f3-4068-9322-ccdb97c8470d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.545414] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Releasing lock "refresh_cache-e6af4651-9f3a-4ce0-add8-06f1cfef255f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 708.545703] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Instance network_info: |[{"id": "6be6bb54-d3f3-4068-9322-ccdb97c8470d", "address": "fa:16:3e:a2:b6:05", "network": {"id": "af150ca9-547c-4640-ae29-015a75200623", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-81489590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3687fbff426473b9d6135208426b66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6be6bb54-d3", "ovs_interfaceid": "6be6bb54-d3f3-4068-9322-ccdb97c8470d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 708.545980] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5269f00b-1102-f75d-0cc8-af1319f69d29, 'name': SearchDatastore_Task, 'duration_secs': 0.018207} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.556504] env[62814]: DEBUG oslo_concurrency.lockutils [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] Acquired lock "refresh_cache-e6af4651-9f3a-4ce0-add8-06f1cfef255f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 708.556702] env[62814]: DEBUG nova.network.neutron [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Refreshing network info cache for port 6be6bb54-d3f3-4068-9322-ccdb97c8470d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 708.558391] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:b6:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c894ab55-c869-4530-9702-cb46d173ce94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6be6bb54-d3f3-4068-9322-ccdb97c8470d', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 708.566777] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Creating folder: Project (e3687fbff426473b9d6135208426b66e). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 708.569494] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 708.569914] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52ea45da-02d3-4ae8-818b-494b97cb98dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.571764] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5167a06b-1b23-4149-a1d6-37d939af323b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.590282] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 708.590282] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fbbb3c-4008-8e88-292d-f389bcc54aa1" [ 708.590282] env[62814]: _type = "Task" [ 708.590282] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.590476] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293424, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.590818] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293422, 'name': Rename_Task, 'duration_secs': 1.309874} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.592362] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 708.592627] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Created folder: Project (e3687fbff426473b9d6135208426b66e) in parent group-v845547. [ 708.592797] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Creating folder: Instances. Parent ref: group-v845619. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 708.595824] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66e889b1-8b47-40d3-add8-54da408e502a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.597755] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-42210f57-d09e-49bf-a33e-82b7a2cf31f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.605889] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fbbb3c-4008-8e88-292d-f389bcc54aa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.609950] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 708.609950] env[62814]: value = "task-4293426" [ 708.609950] env[62814]: _type = "Task" [ 708.609950] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.611491] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Created folder: Instances in parent group-v845619. [ 708.611712] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 708.615059] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 708.615059] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cbec0cb1-cbd5-4f4f-8d27-4ff38d077df2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.634129] env[62814]: DEBUG nova.network.neutron [req-a1bab209-d973-4e8a-9a9f-98bbd8957c24 req-95bb2660-8631-425d-ad49-50c9e07c99b7 service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Updated VIF entry in instance network info cache for port 7aee2670-a9fd-4740-9463-7492279dcfc6. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 708.634401] env[62814]: DEBUG nova.network.neutron [req-a1bab209-d973-4e8a-9a9f-98bbd8957c24 req-95bb2660-8631-425d-ad49-50c9e07c99b7 service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Updating instance_info_cache with network_info: [{"id": "7aee2670-a9fd-4740-9463-7492279dcfc6", "address": "fa:16:3e:20:98:ff", "network": {"id": "a1bdb789-23d9-4db4-b159-376e397cad6c", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-687914596-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b776e7b42ebb42f09820b9005d04049a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee555dfd-3d1a-4220-89cd-ffba64e4acf0", "external-id": "nsx-vlan-transportzone-88", "segmentation_id": 88, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7aee2670-a9", "ovs_interfaceid": "7aee2670-a9fd-4740-9463-7492279dcfc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.639662] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293426, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.641247] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 708.641247] env[62814]: value = "task-4293428" [ 708.641247] env[62814]: _type = "Task" [ 708.641247] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.652206] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293428, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.850796] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c5277e-2298-467f-934f-c0a59820e4d8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.859064] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d7f4ef-6be8-4a0a-8beb-fb9a7391196b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.890791] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc2b9d1-de39-4428-a468-62ba4f6d319e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.898440] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867e2f0c-eae4-45b0-ad09-ca882171638f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.912181] env[62814]: DEBUG nova.compute.provider_tree [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.056644] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293424, 'name': Rename_Task, 'duration_secs': 1.142027} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.056932] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 709.057274] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-328ea4e7-76b7-4eaf-9a8b-18eb56babf8d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.065840] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 709.065840] env[62814]: value = "task-4293429" [ 709.065840] env[62814]: _type = "Task" [ 709.065840] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.073534] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.104210] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fbbb3c-4008-8e88-292d-f389bcc54aa1, 'name': SearchDatastore_Task, 'duration_secs': 0.025781} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.104476] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 709.104738] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] f5ad9d70-75fb-4881-8853-5ede4d0903f2/f5ad9d70-75fb-4881-8853-5ede4d0903f2.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 709.105028] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ffdd0848-fffe-47b2-bc24-26d50aa993a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.112561] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 709.112561] env[62814]: value = "task-4293430" [ 709.112561] env[62814]: _type = "Task" [ 709.112561] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.123803] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293426, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.126935] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293430, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.142308] env[62814]: DEBUG oslo_concurrency.lockutils [req-a1bab209-d973-4e8a-9a9f-98bbd8957c24 req-95bb2660-8631-425d-ad49-50c9e07c99b7 service nova] Releasing lock "refresh_cache-8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 709.151811] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293428, 'name': CreateVM_Task, 'duration_secs': 0.328656} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.152031] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 709.152700] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.152935] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 709.153277] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 709.153543] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c671aa4-957c-4ca9-987b-9f4225da7403 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.158144] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 709.158144] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52da0f50-0f24-b498-50f2-8ff60d16be60" [ 709.158144] env[62814]: _type = "Task" [ 709.158144] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.173554] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52da0f50-0f24-b498-50f2-8ff60d16be60, 'name': SearchDatastore_Task, 'duration_secs': 0.008921} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.174276] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 709.174365] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 709.174627] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.174777] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 709.174958] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 709.175246] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd2752a2-b7fb-4044-be19-285bfb9e4a34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.183565] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 709.183752] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 709.184478] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-836ab8da-7679-4c73-94ca-e671fc5ab61c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.190075] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 709.190075] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522c651f-a396-8126-4bcd-0862eee9a3c8" [ 709.190075] env[62814]: _type = "Task" [ 709.190075] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.201087] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522c651f-a396-8126-4bcd-0862eee9a3c8, 'name': SearchDatastore_Task, 'duration_secs': 0.00823} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.202382] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41a861a4-a18b-4c0d-8058-009b62150158 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.207843] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 709.207843] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d912a8-a900-cd13-62cc-ee9ab5e9d560" [ 709.207843] env[62814]: _type = "Task" [ 709.207843] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.215818] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d912a8-a900-cd13-62cc-ee9ab5e9d560, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.330469] env[62814]: DEBUG nova.network.neutron [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Updated VIF entry in instance network info cache for port 6be6bb54-d3f3-4068-9322-ccdb97c8470d. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 709.331372] env[62814]: DEBUG nova.network.neutron [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Updating instance_info_cache with network_info: [{"id": "6be6bb54-d3f3-4068-9322-ccdb97c8470d", "address": "fa:16:3e:a2:b6:05", "network": {"id": "af150ca9-547c-4640-ae29-015a75200623", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-81489590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3687fbff426473b9d6135208426b66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6be6bb54-d3", "ovs_interfaceid": "6be6bb54-d3f3-4068-9322-ccdb97c8470d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.419026] env[62814]: DEBUG nova.scheduler.client.report [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 709.578517] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293429, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.627220] env[62814]: DEBUG oslo_vmware.api [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293426, 'name': PowerOnVM_Task, 'duration_secs': 0.654987} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.632157] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 709.632488] env[62814]: INFO nova.compute.manager [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Took 11.78 seconds to spawn the instance on the hypervisor. [ 709.632806] env[62814]: DEBUG nova.compute.manager [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 709.633289] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293430, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.634448] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2417e75-d3e0-4a67-9754-af580de911e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.718090] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d912a8-a900-cd13-62cc-ee9ab5e9d560, 'name': SearchDatastore_Task, 'duration_secs': 0.012199} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.718090] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 709.718322] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] e6af4651-9f3a-4ce0-add8-06f1cfef255f/e6af4651-9f3a-4ce0-add8-06f1cfef255f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 709.718570] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ddcf9af-1918-487a-b22d-cb2b2d13196d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.724804] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 709.724804] env[62814]: value = "task-4293431" [ 709.724804] env[62814]: _type = "Task" [ 709.724804] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.732260] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.834232] env[62814]: DEBUG oslo_concurrency.lockutils [req-a7d56f32-d74a-450d-97e1-6426f71f2176 req-0ce0e5c5-a3d4-44a4-bb06-18c4eb76371a service nova] Releasing lock "refresh_cache-e6af4651-9f3a-4ce0-add8-06f1cfef255f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 709.924614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.418s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.925191] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 709.927842] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.979s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.928560] env[62814]: DEBUG nova.objects.instance [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lazy-loading 'resources' on Instance uuid 6976b964-a8d3-4886-8aac-8d513e721018 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 710.076616] env[62814]: DEBUG oslo_vmware.api [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293429, 'name': PowerOnVM_Task, 'duration_secs': 0.599083} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.076908] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 710.077131] env[62814]: INFO nova.compute.manager [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Took 9.41 seconds to spawn the instance on the hypervisor. [ 710.077314] env[62814]: DEBUG nova.compute.manager [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 710.078080] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a5da9d-29e7-4a39-96bd-bd6aa0e611a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.126499] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293430, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.565306} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.127465] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] f5ad9d70-75fb-4881-8853-5ede4d0903f2/f5ad9d70-75fb-4881-8853-5ede4d0903f2.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 710.127806] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 710.129724] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45026af9-a594-41fd-af72-1f42bf2df8b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.135407] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 710.135407] env[62814]: value = "task-4293432" [ 710.135407] env[62814]: _type = "Task" [ 710.135407] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.144553] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293432, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.156205] env[62814]: INFO nova.compute.manager [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Took 48.36 seconds to build instance. [ 710.235413] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293431, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.432055] env[62814]: DEBUG nova.compute.utils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 710.436539] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 710.436539] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.479750] env[62814]: DEBUG nova.policy [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a2e083ae164419fa78addf5da979950', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3687fbff426473b9d6135208426b66e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 710.600654] env[62814]: INFO nova.compute.manager [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Took 45.55 seconds to build instance. [ 710.644981] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293432, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085324} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.645416] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 710.648727] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35dc2ec-42a5-4c66-9549-70e89d9cf639 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.665251] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fdc2516-bf7e-4463-9893-4d6437f347e0 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "4ed66b36-b6c6-4673-9c03-169a01134574" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.509s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.673632] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Reconfiguring VM instance instance-00000018 to attach disk [datastore2] f5ad9d70-75fb-4881-8853-5ede4d0903f2/f5ad9d70-75fb-4881-8853-5ede4d0903f2.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 710.677886] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4033614-8d8e-4212-bba8-d6c2f81ca99b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.699151] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 710.699151] env[62814]: value = "task-4293433" [ 710.699151] env[62814]: _type = "Task" [ 710.699151] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.713016] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293433, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.737939] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293431, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.761654} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.738150] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] e6af4651-9f3a-4ce0-add8-06f1cfef255f/e6af4651-9f3a-4ce0-add8-06f1cfef255f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 710.738380] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 710.738642] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-141772a0-3b94-4992-9f3c-f99a870a2214 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.752021] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 710.752021] env[62814]: value = "task-4293434" [ 710.752021] env[62814]: _type = "Task" [ 710.752021] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.758929] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293434, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.907102] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Successfully created port: 33a51a23-d1ad-456b-9a36-3c0a6df8f966 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.936508] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 711.056107] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea982ab-f60a-4ce7-b0eb-11c4e1393441 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.062455] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581c6219-d8ba-40c5-947e-e8553264e7d1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.095640] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5b164c-39d4-4f76-8402-ab045d254846 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.104508] env[62814]: DEBUG oslo_concurrency.lockutils [None req-57b6d54a-6f49-4c7c-b562-359c368cef64 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "29e69c7c-08f7-4da4-9509-02a94bf971bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.401s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.107107] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c82c07d-ddbb-4612-9989-f3ceed24b096 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.122520] env[62814]: DEBUG nova.compute.provider_tree [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.193707] env[62814]: DEBUG nova.compute.manager [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 711.210285] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293433, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.258231] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293434, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.220852} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.258498] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 711.259325] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1021759-8c55-421f-adbb-476df71967b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.282038] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Reconfiguring VM instance instance-00000019 to attach disk [datastore2] e6af4651-9f3a-4ce0-add8-06f1cfef255f/e6af4651-9f3a-4ce0-add8-06f1cfef255f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 711.282400] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-877c46c8-b3b9-421d-86cf-d29694cc61f2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.305022] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 711.305022] env[62814]: value = "task-4293435" [ 711.305022] env[62814]: _type = "Task" [ 711.305022] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.313780] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293435, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.613041] env[62814]: DEBUG nova.compute.manager [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 711.626491] env[62814]: DEBUG nova.scheduler.client.report [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 711.713388] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293433, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.722962] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 711.819065] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293435, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.948514] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 711.974547] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 711.974796] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.975013] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 711.975262] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.976021] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 711.976021] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 711.976021] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 711.976021] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 711.976232] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 711.976232] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 711.976382] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 711.977443] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c20557d1-0f02-4cd8-b93c-07bbd0295eda {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.985454] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19438566-26fe-47c6-a424-b9e4b1139daa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.133609] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.205s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.135857] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.497s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.137834] env[62814]: INFO nova.compute.claims [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.141900] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 712.163031] env[62814]: INFO nova.scheduler.client.report [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Deleted allocations for instance 6976b964-a8d3-4886-8aac-8d513e721018 [ 712.216659] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293433, 'name': ReconfigVM_Task, 'duration_secs': 1.150123} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.216972] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Reconfigured VM instance instance-00000018 to attach disk [datastore2] f5ad9d70-75fb-4881-8853-5ede4d0903f2/f5ad9d70-75fb-4881-8853-5ede4d0903f2.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 712.218806] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8b9eafc-21ab-4a15-8f57-aedefc3a3278 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.225668] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 712.225668] env[62814]: value = "task-4293436" [ 712.225668] env[62814]: _type = "Task" [ 712.225668] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.235871] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293436, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.316640] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293435, 'name': ReconfigVM_Task, 'duration_secs': 0.846571} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.316975] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Reconfigured VM instance instance-00000019 to attach disk [datastore2] e6af4651-9f3a-4ce0-add8-06f1cfef255f/e6af4651-9f3a-4ce0-add8-06f1cfef255f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 712.317635] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1453e43e-783f-4346-b9bb-289f5269909a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.323880] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 712.323880] env[62814]: value = "task-4293437" [ 712.323880] env[62814]: _type = "Task" [ 712.323880] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.334744] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293437, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.578518] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Successfully updated port: 33a51a23-d1ad-456b-9a36-3c0a6df8f966 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 712.619851] env[62814]: DEBUG nova.compute.manager [req-0943087c-3c00-42c2-bc43-4228e9b802f0 req-c2215e5e-bd2c-4a39-ac60-19106dce9811 service nova] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Received event network-vif-plugged-33a51a23-d1ad-456b-9a36-3c0a6df8f966 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 712.620125] env[62814]: DEBUG oslo_concurrency.lockutils [req-0943087c-3c00-42c2-bc43-4228e9b802f0 req-c2215e5e-bd2c-4a39-ac60-19106dce9811 service nova] Acquiring lock "64a475e4-6713-408b-a63a-a43b5fed5ec8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 712.620679] env[62814]: DEBUG oslo_concurrency.lockutils [req-0943087c-3c00-42c2-bc43-4228e9b802f0 req-c2215e5e-bd2c-4a39-ac60-19106dce9811 service nova] Lock "64a475e4-6713-408b-a63a-a43b5fed5ec8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.620679] env[62814]: DEBUG oslo_concurrency.lockutils [req-0943087c-3c00-42c2-bc43-4228e9b802f0 req-c2215e5e-bd2c-4a39-ac60-19106dce9811 service nova] Lock "64a475e4-6713-408b-a63a-a43b5fed5ec8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.620765] env[62814]: DEBUG nova.compute.manager [req-0943087c-3c00-42c2-bc43-4228e9b802f0 req-c2215e5e-bd2c-4a39-ac60-19106dce9811 service nova] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] No waiting events found dispatching network-vif-plugged-33a51a23-d1ad-456b-9a36-3c0a6df8f966 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 712.620926] env[62814]: WARNING nova.compute.manager [req-0943087c-3c00-42c2-bc43-4228e9b802f0 req-c2215e5e-bd2c-4a39-ac60-19106dce9811 service nova] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Received unexpected event network-vif-plugged-33a51a23-d1ad-456b-9a36-3c0a6df8f966 for instance with vm_state building and task_state spawning. [ 712.672075] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8cf4a4a1-e419-4dc2-acb0-1b254d1f2690 tempest-AttachInterfacesV270Test-494041876 tempest-AttachInterfacesV270Test-494041876-project-member] Lock "6976b964-a8d3-4886-8aac-8d513e721018" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.150s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.736246] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293436, 'name': Rename_Task, 'duration_secs': 0.150941} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.736406] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 712.736654] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49155e18-6bbd-4597-b5f3-30124ed5237a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.746078] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 712.746078] env[62814]: value = "task-4293438" [ 712.746078] env[62814]: _type = "Task" [ 712.746078] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.756243] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293438, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.835601] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293437, 'name': Rename_Task, 'duration_secs': 0.202243} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.835944] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 712.836254] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1fd75b98-8dd6-4677-af19-59be02e8be89 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.842172] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 712.842172] env[62814]: value = "task-4293439" [ 712.842172] env[62814]: _type = "Task" [ 712.842172] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.850812] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293439, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.083490] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "refresh_cache-64a475e4-6713-408b-a63a-a43b5fed5ec8" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.083647] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired lock "refresh_cache-64a475e4-6713-408b-a63a-a43b5fed5ec8" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.083803] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.255644] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293438, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.351726] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293439, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.599475] env[62814]: DEBUG nova.compute.manager [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 713.600551] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7831c887-7259-4bce-bb47-153a9a1a5bf9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.636694] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.684150] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0543938-9e3a-49a5-a606-adb2cc0f22fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.692950] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c992f469-5dbe-48df-b095-65c25a521918 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.730040] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba3b79b-e15b-4b77-bea7-6dd2c032d4c1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.736357] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96c591f-c713-4a03-9efe-c278a2d28241 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.750690] env[62814]: DEBUG nova.compute.provider_tree [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.761351] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293438, 'name': PowerOnVM_Task} progress is 91%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.852797] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293439, 'name': PowerOnVM_Task, 'duration_secs': 0.608069} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.855836] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 713.855836] env[62814]: INFO nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Took 7.59 seconds to spawn the instance on the hypervisor. [ 713.855836] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 713.856393] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab817902-af92-4a3f-b02c-2fe7c8586cf7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.920420] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Updating instance_info_cache with network_info: [{"id": "33a51a23-d1ad-456b-9a36-3c0a6df8f966", "address": "fa:16:3e:78:f1:4f", "network": {"id": "af150ca9-547c-4640-ae29-015a75200623", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-81489590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3687fbff426473b9d6135208426b66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33a51a23-d1", "ovs_interfaceid": "33a51a23-d1ad-456b-9a36-3c0a6df8f966", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.117596] env[62814]: INFO nova.compute.manager [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] instance snapshotting [ 714.121175] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0852a849-2d26-4a6e-9a28-7f91d6f24f81 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.141035] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dbd096-33e8-4ae7-a3de-9d346e97ed04 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.257063] env[62814]: DEBUG nova.scheduler.client.report [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 714.266776] env[62814]: DEBUG oslo_vmware.api [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293438, 'name': PowerOnVM_Task, 'duration_secs': 1.01972} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.268567] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 714.268567] env[62814]: INFO nova.compute.manager [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Took 10.77 seconds to spawn the instance on the hypervisor. [ 714.268567] env[62814]: DEBUG nova.compute.manager [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 714.269144] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76dc725b-1848-4838-bb26-81843991fcca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.383153] env[62814]: INFO nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Took 47.43 seconds to build instance. [ 714.424352] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Releasing lock "refresh_cache-64a475e4-6713-408b-a63a-a43b5fed5ec8" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 714.424708] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Instance network_info: |[{"id": "33a51a23-d1ad-456b-9a36-3c0a6df8f966", "address": "fa:16:3e:78:f1:4f", "network": {"id": "af150ca9-547c-4640-ae29-015a75200623", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-81489590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3687fbff426473b9d6135208426b66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33a51a23-d1", "ovs_interfaceid": "33a51a23-d1ad-456b-9a36-3c0a6df8f966", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 714.425132] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:f1:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c894ab55-c869-4530-9702-cb46d173ce94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33a51a23-d1ad-456b-9a36-3c0a6df8f966', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 714.435558] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 714.435816] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 714.436317] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-193f8f80-3bbe-4bd5-92bc-9201d04e7a3c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.457754] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 714.457754] env[62814]: value = "task-4293440" [ 714.457754] env[62814]: _type = "Task" [ 714.457754] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.466506] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293440, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.651714] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 714.653515] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1ae7c26d-39d5-4fdd-8b33-b66b51c1de3f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.656359] env[62814]: DEBUG nova.compute.manager [req-dee55489-a9b6-45d3-82a7-15bff93b7dd1 req-b1ae9877-a92a-44fd-9f93-4a5b90c6cb7f service nova] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Received event network-changed-33a51a23-d1ad-456b-9a36-3c0a6df8f966 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 714.656547] env[62814]: DEBUG nova.compute.manager [req-dee55489-a9b6-45d3-82a7-15bff93b7dd1 req-b1ae9877-a92a-44fd-9f93-4a5b90c6cb7f service nova] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Refreshing instance network info cache due to event network-changed-33a51a23-d1ad-456b-9a36-3c0a6df8f966. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 714.656768] env[62814]: DEBUG oslo_concurrency.lockutils [req-dee55489-a9b6-45d3-82a7-15bff93b7dd1 req-b1ae9877-a92a-44fd-9f93-4a5b90c6cb7f service nova] Acquiring lock "refresh_cache-64a475e4-6713-408b-a63a-a43b5fed5ec8" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.656923] env[62814]: DEBUG oslo_concurrency.lockutils [req-dee55489-a9b6-45d3-82a7-15bff93b7dd1 req-b1ae9877-a92a-44fd-9f93-4a5b90c6cb7f service nova] Acquired lock "refresh_cache-64a475e4-6713-408b-a63a-a43b5fed5ec8" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 714.657111] env[62814]: DEBUG nova.network.neutron [req-dee55489-a9b6-45d3-82a7-15bff93b7dd1 req-b1ae9877-a92a-44fd-9f93-4a5b90c6cb7f service nova] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Refreshing network info cache for port 33a51a23-d1ad-456b-9a36-3c0a6df8f966 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 714.666374] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 714.666374] env[62814]: value = "task-4293441" [ 714.666374] env[62814]: _type = "Task" [ 714.666374] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.674952] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293441, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.772239] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.636s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.772944] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 714.776229] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.508s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.778051] env[62814]: INFO nova.compute.claims [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.801598] env[62814]: INFO nova.compute.manager [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Took 48.33 seconds to build instance. [ 714.887789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.866s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.967805] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293440, 'name': CreateVM_Task, 'duration_secs': 0.497104} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.967973] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 714.969173] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.969348] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 714.969650] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 714.969905] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1496f419-badb-4b11-8f89-64ddde2c9820 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.975318] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 714.975318] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c1cad2-bcb2-18a9-1c97-0defd573eaf7" [ 714.975318] env[62814]: _type = "Task" [ 714.975318] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.984409] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c1cad2-bcb2-18a9-1c97-0defd573eaf7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.176448] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293441, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.290295] env[62814]: DEBUG nova.compute.utils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 715.299892] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 715.302021] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 715.305703] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2d602646-cf87-4706-83fb-8cf5b351a81a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.693s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 715.360128] env[62814]: DEBUG nova.policy [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a2e083ae164419fa78addf5da979950', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e3687fbff426473b9d6135208426b66e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 715.390063] env[62814]: DEBUG nova.compute.manager [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 715.485864] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c1cad2-bcb2-18a9-1c97-0defd573eaf7, 'name': SearchDatastore_Task, 'duration_secs': 0.014934} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.486195] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 715.486430] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 715.486660] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.486814] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 715.487035] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 715.488271] env[62814]: DEBUG nova.network.neutron [req-dee55489-a9b6-45d3-82a7-15bff93b7dd1 req-b1ae9877-a92a-44fd-9f93-4a5b90c6cb7f service nova] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Updated VIF entry in instance network info cache for port 33a51a23-d1ad-456b-9a36-3c0a6df8f966. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 715.488737] env[62814]: DEBUG nova.network.neutron [req-dee55489-a9b6-45d3-82a7-15bff93b7dd1 req-b1ae9877-a92a-44fd-9f93-4a5b90c6cb7f service nova] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Updating instance_info_cache with network_info: [{"id": "33a51a23-d1ad-456b-9a36-3c0a6df8f966", "address": "fa:16:3e:78:f1:4f", "network": {"id": "af150ca9-547c-4640-ae29-015a75200623", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-81489590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3687fbff426473b9d6135208426b66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33a51a23-d1", "ovs_interfaceid": "33a51a23-d1ad-456b-9a36-3c0a6df8f966", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.491043] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-17d4eff9-4fcd-4b27-b441-c22b26dffc3f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.503088] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 715.503309] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 715.504494] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2fbc7dd-9c15-4ac3-abe7-c83afcb102a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.512413] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 715.512413] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c2ad70-88d5-c362-56b4-685f529c3db2" [ 715.512413] env[62814]: _type = "Task" [ 715.512413] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.522715] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c2ad70-88d5-c362-56b4-685f529c3db2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.678094] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293441, 'name': CreateSnapshot_Task, 'duration_secs': 0.766388} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.678381] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 715.679249] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7fbed35-b5af-422b-835b-7b7dd99bab90 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.760679] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Successfully created port: cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.800998] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 715.813906] env[62814]: DEBUG nova.compute.manager [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 715.912558] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 715.993952] env[62814]: DEBUG oslo_concurrency.lockutils [req-dee55489-a9b6-45d3-82a7-15bff93b7dd1 req-b1ae9877-a92a-44fd-9f93-4a5b90c6cb7f service nova] Releasing lock "refresh_cache-64a475e4-6713-408b-a63a-a43b5fed5ec8" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.025856] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c2ad70-88d5-c362-56b4-685f529c3db2, 'name': SearchDatastore_Task, 'duration_secs': 0.028621} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.030135] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b697d8f-cd7c-4d3f-93e8-65a2b31858df {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.037776] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 716.037776] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523fe773-ffc9-d5f4-50bd-a46ce88857a8" [ 716.037776] env[62814]: _type = "Task" [ 716.037776] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.048148] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523fe773-ffc9-d5f4-50bd-a46ce88857a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.205134] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 716.205815] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-207f5177-1c9d-4831-90b6-6ba2ef7bad45 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.215357] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 716.215357] env[62814]: value = "task-4293442" [ 716.215357] env[62814]: _type = "Task" [ 716.215357] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.228187] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293442, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.350247] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.457750] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 716.457750] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 716.511015] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-038a1e54-986e-426b-ab61-e828c50df070 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.521035] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6801bcbb-626c-4acd-a385-cde9aa1e4720 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.558804] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa41e44-59b4-483f-8e7b-2c118d0c2d13 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.570188] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ccec986-9890-4a49-b5f1-dd7f52ff5a4c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.574543] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523fe773-ffc9-d5f4-50bd-a46ce88857a8, 'name': SearchDatastore_Task, 'duration_secs': 0.014324} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.574827] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 716.575174] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 64a475e4-6713-408b-a63a-a43b5fed5ec8/64a475e4-6713-408b-a63a-a43b5fed5ec8.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 716.576505] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0cb8360a-6753-4517-88a7-9201bef72d79 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.588832] env[62814]: DEBUG nova.compute.provider_tree [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.596892] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 716.596892] env[62814]: value = "task-4293443" [ 716.596892] env[62814]: _type = "Task" [ 716.596892] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.606941] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293443, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.727967] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293442, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.826934] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 716.862081] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 716.862346] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.862861] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 716.862861] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.862861] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 716.862985] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 716.863415] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 716.863818] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 716.864143] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 716.864473] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 716.864797] env[62814]: DEBUG nova.virt.hardware [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 716.866654] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684641fa-3967-4a25-a7eb-8a1fdee3f799 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.877555] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842f7c40-1e79-4465-b3f5-2252337fc704 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.093254] env[62814]: DEBUG nova.scheduler.client.report [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 717.111066] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293443, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.226919] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293442, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.604089] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.828s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 717.604842] env[62814]: DEBUG nova.compute.manager [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 717.613244] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.505s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 717.613244] env[62814]: DEBUG nova.objects.instance [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lazy-loading 'resources' on Instance uuid 62bc755d-4f96-4486-884b-0d0c337267aa {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 717.613417] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293443, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659021} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.614117] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 64a475e4-6713-408b-a63a-a43b5fed5ec8/64a475e4-6713-408b-a63a-a43b5fed5ec8.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 717.616517] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.616800] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-376849c7-95a3-4fba-a847-4655741f5062 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.626235] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 717.626235] env[62814]: value = "task-4293444" [ 717.626235] env[62814]: _type = "Task" [ 717.626235] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.636101] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293444, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.729257] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293442, 'name': CloneVM_Task} progress is 95%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.951024] env[62814]: DEBUG nova.compute.manager [req-f8a316ba-8f0f-43fd-9b74-fd0940627a00 req-55f8cbbd-1de5-4857-ad7f-a4e484981be8 service nova] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Received event network-vif-plugged-cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 717.951495] env[62814]: DEBUG oslo_concurrency.lockutils [req-f8a316ba-8f0f-43fd-9b74-fd0940627a00 req-55f8cbbd-1de5-4857-ad7f-a4e484981be8 service nova] Acquiring lock "961f313b-b43f-4531-8a4b-0a39421d6a34-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 717.953977] env[62814]: DEBUG oslo_concurrency.lockutils [req-f8a316ba-8f0f-43fd-9b74-fd0940627a00 req-55f8cbbd-1de5-4857-ad7f-a4e484981be8 service nova] Lock "961f313b-b43f-4531-8a4b-0a39421d6a34-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 717.953977] env[62814]: DEBUG oslo_concurrency.lockutils [req-f8a316ba-8f0f-43fd-9b74-fd0940627a00 req-55f8cbbd-1de5-4857-ad7f-a4e484981be8 service nova] Lock "961f313b-b43f-4531-8a4b-0a39421d6a34-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 717.954142] env[62814]: DEBUG nova.compute.manager [req-f8a316ba-8f0f-43fd-9b74-fd0940627a00 req-55f8cbbd-1de5-4857-ad7f-a4e484981be8 service nova] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] No waiting events found dispatching network-vif-plugged-cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 717.954737] env[62814]: WARNING nova.compute.manager [req-f8a316ba-8f0f-43fd-9b74-fd0940627a00 req-55f8cbbd-1de5-4857-ad7f-a4e484981be8 service nova] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Received unexpected event network-vif-plugged-cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b for instance with vm_state building and task_state spawning. [ 718.059045] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Successfully updated port: cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 718.115717] env[62814]: DEBUG nova.compute.utils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 718.121831] env[62814]: DEBUG nova.compute.manager [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 718.121831] env[62814]: DEBUG nova.network.neutron [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 718.136030] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293444, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.176024} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.136268] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 718.137646] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-857f45a0-ac85-47d6-a396-12a69222792e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.161699] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] 64a475e4-6713-408b-a63a-a43b5fed5ec8/64a475e4-6713-408b-a63a-a43b5fed5ec8.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 718.167238] env[62814]: DEBUG nova.policy [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3778ab0909754e4bb077396f985b0630', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '023cfd6e178c4d77b5e662340924e101', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 718.167238] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5a152d7-3518-4fde-aa32-009a920fafe8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.186686] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 718.186686] env[62814]: value = "task-4293445" [ 718.186686] env[62814]: _type = "Task" [ 718.186686] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.195043] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293445, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.231136] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293442, 'name': CloneVM_Task, 'duration_secs': 1.834289} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.231441] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Created linked-clone VM from snapshot [ 718.232223] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c271d4-5430-4ab1-a669-9a0247ed5c72 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.243253] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Uploading image f971bd36-a39d-4845-8e7e-dfce875c390a {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 718.257608] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 718.257878] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4d324eb2-8b06-4f27-a587-1631d39f5f62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.267993] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 718.267993] env[62814]: value = "task-4293446" [ 718.267993] env[62814]: _type = "Task" [ 718.267993] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.278386] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293446, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.565408] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "refresh_cache-961f313b-b43f-4531-8a4b-0a39421d6a34" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.565609] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired lock "refresh_cache-961f313b-b43f-4531-8a4b-0a39421d6a34" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 718.565838] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 718.607023] env[62814]: DEBUG nova.network.neutron [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Successfully created port: b8697072-9a1f-46f2-9da3-c372d47e8385 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 718.623543] env[62814]: DEBUG nova.compute.manager [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 718.704559] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293445, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.771619] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656ba1fb-3494-427a-8970-f816dbb3a7a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.791414] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d38284-0d3a-4127-9328-8e067af0db38 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.797010] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293446, 'name': Destroy_Task, 'duration_secs': 0.298421} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.797318] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Destroyed the VM [ 718.797581] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 718.798348] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-df763a05-082b-4456-b20f-3d9d727be357 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.830974] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b561ab39-0d7c-4ee5-ae6f-2e54df9f45ef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.835313] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 718.835313] env[62814]: value = "task-4293447" [ 718.835313] env[62814]: _type = "Task" [ 718.835313] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.843960] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3ef1cd0-9be3-4a8c-a3dc-d38fb855c985 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.851767] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293447, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.865427] env[62814]: DEBUG nova.compute.provider_tree [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.127682] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.201186] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293445, 'name': ReconfigVM_Task, 'duration_secs': 0.56473} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.201186] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Reconfigured VM instance instance-0000001a to attach disk [datastore2] 64a475e4-6713-408b-a63a-a43b5fed5ec8/64a475e4-6713-408b-a63a-a43b5fed5ec8.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 719.201629] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-064a0aba-03c5-4d0e-a45d-17d2369fdcd9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.210428] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 719.210428] env[62814]: value = "task-4293448" [ 719.210428] env[62814]: _type = "Task" [ 719.210428] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.221629] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293448, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.309727] env[62814]: DEBUG nova.network.neutron [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Updating instance_info_cache with network_info: [{"id": "cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b", "address": "fa:16:3e:23:07:87", "network": {"id": "af150ca9-547c-4640-ae29-015a75200623", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-81489590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3687fbff426473b9d6135208426b66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb1f398a-cf", "ovs_interfaceid": "cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.347275] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293447, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.369923] env[62814]: DEBUG nova.scheduler.client.report [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 719.639507] env[62814]: DEBUG nova.compute.manager [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 719.673533] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 719.673809] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 719.673962] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 719.674241] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 719.674406] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 719.674566] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 719.674787] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 719.674964] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 719.675223] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 719.675414] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 719.675607] env[62814]: DEBUG nova.virt.hardware [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 719.676514] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1ca622-328c-438f-91b7-8aec151e0133 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.685994] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85dc6a1-b05f-49d1-a374-6a4d5c16cc9c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.719084] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293448, 'name': Rename_Task, 'duration_secs': 0.205723} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.719511] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 719.719759] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8149dca6-e44c-4e00-8ce9-99d6ab3b67df {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.727965] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 719.727965] env[62814]: value = "task-4293449" [ 719.727965] env[62814]: _type = "Task" [ 719.727965] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.736218] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293449, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.815046] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Releasing lock "refresh_cache-961f313b-b43f-4531-8a4b-0a39421d6a34" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 719.815493] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Instance network_info: |[{"id": "cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b", "address": "fa:16:3e:23:07:87", "network": {"id": "af150ca9-547c-4640-ae29-015a75200623", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-81489590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3687fbff426473b9d6135208426b66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb1f398a-cf", "ovs_interfaceid": "cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 719.815801] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:07:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c894ab55-c869-4530-9702-cb46d173ce94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 719.829275] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 719.829851] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 719.829851] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b210b6c-0c5e-43a9-9a1d-66f57df9c80c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.862833] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293447, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.864034] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 719.864034] env[62814]: value = "task-4293450" [ 719.864034] env[62814]: _type = "Task" [ 719.864034] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.874703] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293450, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.875696] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.265s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 719.878815] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.632s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 719.881071] env[62814]: INFO nova.compute.claims [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.905690] env[62814]: INFO nova.scheduler.client.report [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted allocations for instance 62bc755d-4f96-4486-884b-0d0c337267aa [ 720.013133] env[62814]: DEBUG nova.compute.manager [req-bf8cb80e-0be7-4de9-8f56-674e471acbd8 req-9258a533-324a-4f4b-b53a-5cb48cea0227 service nova] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Received event network-changed-cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 720.013364] env[62814]: DEBUG nova.compute.manager [req-bf8cb80e-0be7-4de9-8f56-674e471acbd8 req-9258a533-324a-4f4b-b53a-5cb48cea0227 service nova] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Refreshing instance network info cache due to event network-changed-cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 720.013600] env[62814]: DEBUG oslo_concurrency.lockutils [req-bf8cb80e-0be7-4de9-8f56-674e471acbd8 req-9258a533-324a-4f4b-b53a-5cb48cea0227 service nova] Acquiring lock "refresh_cache-961f313b-b43f-4531-8a4b-0a39421d6a34" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.013795] env[62814]: DEBUG oslo_concurrency.lockutils [req-bf8cb80e-0be7-4de9-8f56-674e471acbd8 req-9258a533-324a-4f4b-b53a-5cb48cea0227 service nova] Acquired lock "refresh_cache-961f313b-b43f-4531-8a4b-0a39421d6a34" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.013956] env[62814]: DEBUG nova.network.neutron [req-bf8cb80e-0be7-4de9-8f56-674e471acbd8 req-9258a533-324a-4f4b-b53a-5cb48cea0227 service nova] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Refreshing network info cache for port cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 720.239865] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293449, 'name': PowerOnVM_Task} progress is 37%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.361633] env[62814]: DEBUG oslo_vmware.api [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293447, 'name': RemoveSnapshot_Task, 'duration_secs': 1.054303} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.362231] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 720.373044] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293450, 'name': CreateVM_Task, 'duration_secs': 0.40415} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.373251] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 720.374010] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.374239] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.374600] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 720.376327] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e53d50c-fd3d-4f5b-af85-f44906706090 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.379979] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 720.379979] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527caa13-5a1b-16f5-1102-b9c1fc37c3d1" [ 720.379979] env[62814]: _type = "Task" [ 720.379979] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.391304] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527caa13-5a1b-16f5-1102-b9c1fc37c3d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.415774] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ac659d03-36c8-4c1c-8b73-1809cdc94d92 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "62bc755d-4f96-4486-884b-0d0c337267aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.875s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.534724] env[62814]: DEBUG nova.network.neutron [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Successfully updated port: b8697072-9a1f-46f2-9da3-c372d47e8385 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 720.740164] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293449, 'name': PowerOnVM_Task, 'duration_secs': 0.954513} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.740382] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 720.740591] env[62814]: INFO nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Took 8.79 seconds to spawn the instance on the hypervisor. [ 720.740771] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 720.741590] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92eb2dc-12ca-4c95-b26e-dbc398810b85 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.866088] env[62814]: DEBUG nova.network.neutron [req-bf8cb80e-0be7-4de9-8f56-674e471acbd8 req-9258a533-324a-4f4b-b53a-5cb48cea0227 service nova] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Updated VIF entry in instance network info cache for port cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 720.866559] env[62814]: DEBUG nova.network.neutron [req-bf8cb80e-0be7-4de9-8f56-674e471acbd8 req-9258a533-324a-4f4b-b53a-5cb48cea0227 service nova] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Updating instance_info_cache with network_info: [{"id": "cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b", "address": "fa:16:3e:23:07:87", "network": {"id": "af150ca9-547c-4640-ae29-015a75200623", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-81489590-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e3687fbff426473b9d6135208426b66e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb1f398a-cf", "ovs_interfaceid": "cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.873990] env[62814]: WARNING nova.compute.manager [None req-cdfb16bc-94d3-4893-910c-9512131e491b tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Image not found during snapshot: nova.exception.ImageNotFound: Image f971bd36-a39d-4845-8e7e-dfce875c390a could not be found. [ 720.891427] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527caa13-5a1b-16f5-1102-b9c1fc37c3d1, 'name': SearchDatastore_Task, 'duration_secs': 0.018438} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.891741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 720.891972] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 720.892242] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.892392] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 720.892569] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 720.895732] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4430f345-c7e0-458f-8f85-a3f7f45c569a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.907772] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 720.907772] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 720.908457] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90ecbb6c-1c2b-49e2-a3c2-554dc2091687 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.915017] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 720.915017] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52169c25-68dd-8a6f-7ce0-fb1ec9c85ca8" [ 720.915017] env[62814]: _type = "Task" [ 720.915017] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.921716] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52169c25-68dd-8a6f-7ce0-fb1ec9c85ca8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.036660] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquiring lock "refresh_cache-004fd137-4902-4313-a6f7-6c83cd76743d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.036831] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquired lock "refresh_cache-004fd137-4902-4313-a6f7-6c83cd76743d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 721.037024] env[62814]: DEBUG nova.network.neutron [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.271254] env[62814]: INFO nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Took 46.28 seconds to build instance. [ 721.376447] env[62814]: DEBUG oslo_concurrency.lockutils [req-bf8cb80e-0be7-4de9-8f56-674e471acbd8 req-9258a533-324a-4f4b-b53a-5cb48cea0227 service nova] Releasing lock "refresh_cache-961f313b-b43f-4531-8a4b-0a39421d6a34" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.425726] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52169c25-68dd-8a6f-7ce0-fb1ec9c85ca8, 'name': SearchDatastore_Task, 'duration_secs': 0.020397} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.429504] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b186a865-cb81-4188-8df1-941c92958af0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.435706] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 721.435706] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5242d8fe-ffd3-c6c6-2096-a924316e3745" [ 721.435706] env[62814]: _type = "Task" [ 721.435706] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.448116] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5242d8fe-ffd3-c6c6-2096-a924316e3745, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.453944] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8751fc9-d787-4f3b-9cb1-7053bd6a0556 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.460694] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b16207e-14b6-47d2-8a8c-0ab7db270a9a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.493725] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33c32e3-9a15-4e04-956f-1015d38bcc09 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.501423] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a2e838-6d7b-437e-bc36-930a68678c6f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.524906] env[62814]: DEBUG nova.compute.provider_tree [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.583972] env[62814]: DEBUG nova.network.neutron [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.623158] env[62814]: DEBUG oslo_concurrency.lockutils [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "29e69c7c-08f7-4da4-9509-02a94bf971bc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.623463] env[62814]: DEBUG oslo_concurrency.lockutils [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "29e69c7c-08f7-4da4-9509-02a94bf971bc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.623675] env[62814]: DEBUG oslo_concurrency.lockutils [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "29e69c7c-08f7-4da4-9509-02a94bf971bc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.623903] env[62814]: DEBUG oslo_concurrency.lockutils [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "29e69c7c-08f7-4da4-9509-02a94bf971bc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.624129] env[62814]: DEBUG oslo_concurrency.lockutils [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "29e69c7c-08f7-4da4-9509-02a94bf971bc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.626444] env[62814]: INFO nova.compute.manager [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Terminating instance [ 721.777713] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "64a475e4-6713-408b-a63a-a43b5fed5ec8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 56.655s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.789548] env[62814]: DEBUG nova.network.neutron [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Updating instance_info_cache with network_info: [{"id": "b8697072-9a1f-46f2-9da3-c372d47e8385", "address": "fa:16:3e:56:08:a5", "network": {"id": "16ac2170-f206-4a40-b1a3-b48b5ea1ecda", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1327582145-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "023cfd6e178c4d77b5e662340924e101", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da", "external-id": "nsx-vlan-transportzone-813", "segmentation_id": 813, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8697072-9a", "ovs_interfaceid": "b8697072-9a1f-46f2-9da3-c372d47e8385", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.800963] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquiring lock "1f40dc62-1a58-4cfb-8785-c37b68747f37" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 721.801243] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lock "1f40dc62-1a58-4cfb-8785-c37b68747f37" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.945858] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5242d8fe-ffd3-c6c6-2096-a924316e3745, 'name': SearchDatastore_Task, 'duration_secs': 0.028389} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.946224] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 721.946401] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 961f313b-b43f-4531-8a4b-0a39421d6a34/961f313b-b43f-4531-8a4b-0a39421d6a34.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 721.946652] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-edf856bf-8f5e-41ee-8997-f4b1b6d7d514 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.955409] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 721.955409] env[62814]: value = "task-4293451" [ 721.955409] env[62814]: _type = "Task" [ 721.955409] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.964099] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293451, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.028421] env[62814]: DEBUG nova.scheduler.client.report [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 722.053113] env[62814]: DEBUG nova.compute.manager [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Received event network-vif-plugged-b8697072-9a1f-46f2-9da3-c372d47e8385 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 722.053113] env[62814]: DEBUG oslo_concurrency.lockutils [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] Acquiring lock "004fd137-4902-4313-a6f7-6c83cd76743d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 722.053113] env[62814]: DEBUG oslo_concurrency.lockutils [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] Lock "004fd137-4902-4313-a6f7-6c83cd76743d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 722.053113] env[62814]: DEBUG oslo_concurrency.lockutils [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] Lock "004fd137-4902-4313-a6f7-6c83cd76743d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 722.053113] env[62814]: DEBUG nova.compute.manager [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] No waiting events found dispatching network-vif-plugged-b8697072-9a1f-46f2-9da3-c372d47e8385 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 722.053616] env[62814]: WARNING nova.compute.manager [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Received unexpected event network-vif-plugged-b8697072-9a1f-46f2-9da3-c372d47e8385 for instance with vm_state building and task_state spawning. [ 722.053616] env[62814]: DEBUG nova.compute.manager [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Received event network-changed-b8697072-9a1f-46f2-9da3-c372d47e8385 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 722.053616] env[62814]: DEBUG nova.compute.manager [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Refreshing instance network info cache due to event network-changed-b8697072-9a1f-46f2-9da3-c372d47e8385. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 722.053616] env[62814]: DEBUG oslo_concurrency.lockutils [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] Acquiring lock "refresh_cache-004fd137-4902-4313-a6f7-6c83cd76743d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.133824] env[62814]: DEBUG nova.compute.manager [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 722.134149] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.135041] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5319c72c-1e8f-4a03-9f5e-9e2ac8933bbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.143427] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 722.143680] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-966b92ab-b6b5-47e6-a859-9489e0595768 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.150231] env[62814]: DEBUG oslo_vmware.api [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 722.150231] env[62814]: value = "task-4293452" [ 722.150231] env[62814]: _type = "Task" [ 722.150231] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.158370] env[62814]: DEBUG oslo_vmware.api [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.281356] env[62814]: DEBUG nova.compute.manager [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 722.292587] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Releasing lock "refresh_cache-004fd137-4902-4313-a6f7-6c83cd76743d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 722.292936] env[62814]: DEBUG nova.compute.manager [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Instance network_info: |[{"id": "b8697072-9a1f-46f2-9da3-c372d47e8385", "address": "fa:16:3e:56:08:a5", "network": {"id": "16ac2170-f206-4a40-b1a3-b48b5ea1ecda", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1327582145-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "023cfd6e178c4d77b5e662340924e101", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da", "external-id": "nsx-vlan-transportzone-813", "segmentation_id": 813, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8697072-9a", "ovs_interfaceid": "b8697072-9a1f-46f2-9da3-c372d47e8385", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 722.293287] env[62814]: DEBUG oslo_concurrency.lockutils [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] Acquired lock "refresh_cache-004fd137-4902-4313-a6f7-6c83cd76743d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 722.293713] env[62814]: DEBUG nova.network.neutron [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Refreshing network info cache for port b8697072-9a1f-46f2-9da3-c372d47e8385 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.294778] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:56:08:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8697072-9a1f-46f2-9da3-c372d47e8385', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 722.304261] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Creating folder: Project (023cfd6e178c4d77b5e662340924e101). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 722.307755] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4deca02-c8e4-4f80-a39f-6e5926d2c41e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.320680] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Created folder: Project (023cfd6e178c4d77b5e662340924e101) in parent group-v845547. [ 722.320891] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Creating folder: Instances. Parent ref: group-v845626. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 722.321180] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-efc38066-0923-45f2-9bbd-6d82bafa73f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.332175] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Created folder: Instances in parent group-v845626. [ 722.332793] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 722.333102] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 722.333928] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-353cfce7-8c9f-4afc-8b25-83ea6a694a30 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.356468] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 722.356468] env[62814]: value = "task-4293455" [ 722.356468] env[62814]: _type = "Task" [ 722.356468] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.365550] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293455, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.472084] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293451, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.534640] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.656s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 722.535235] env[62814]: DEBUG nova.compute.manager [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 722.538019] env[62814]: DEBUG oslo_concurrency.lockutils [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.889s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 722.538934] env[62814]: DEBUG nova.objects.instance [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lazy-loading 'resources' on Instance uuid a36ad785-2f33-4dbc-bc82-ab4a35020b0f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 722.610623] env[62814]: DEBUG nova.network.neutron [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Updated VIF entry in instance network info cache for port b8697072-9a1f-46f2-9da3-c372d47e8385. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 722.611026] env[62814]: DEBUG nova.network.neutron [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Updating instance_info_cache with network_info: [{"id": "b8697072-9a1f-46f2-9da3-c372d47e8385", "address": "fa:16:3e:56:08:a5", "network": {"id": "16ac2170-f206-4a40-b1a3-b48b5ea1ecda", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1327582145-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "023cfd6e178c4d77b5e662340924e101", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37e8d2ee-abfc-42e2-a8fa-ee5447f1f1da", "external-id": "nsx-vlan-transportzone-813", "segmentation_id": 813, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8697072-9a", "ovs_interfaceid": "b8697072-9a1f-46f2-9da3-c372d47e8385", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.663183] env[62814]: DEBUG oslo_vmware.api [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293452, 'name': PowerOffVM_Task, 'duration_secs': 0.330003} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.664643] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 722.664888] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 722.665263] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f175e363-c037-457d-8f37-a02edb6c15f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.753512] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 722.754502] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 722.754502] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Deleting the datastore file [datastore2] 29e69c7c-08f7-4da4-9509-02a94bf971bc {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 722.754502] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b70be4b8-7d01-4246-8d47-6d4f5ca6080b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.763075] env[62814]: DEBUG oslo_vmware.api [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 722.763075] env[62814]: value = "task-4293457" [ 722.763075] env[62814]: _type = "Task" [ 722.763075] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.772270] env[62814]: DEBUG oslo_vmware.api [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293457, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.822535] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 722.869125] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293455, 'name': CreateVM_Task, 'duration_secs': 0.482291} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.869317] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 722.870029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.870194] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 722.870546] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 722.870822] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fee6aef-bb46-4373-9fae-6ca4d9ab70c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.879610] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for the task: (returnval){ [ 722.879610] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5244f567-7ee3-3da6-b526-4ffa9fb8bd7d" [ 722.879610] env[62814]: _type = "Task" [ 722.879610] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.889412] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5244f567-7ee3-3da6-b526-4ffa9fb8bd7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.965230] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293451, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537641} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.965536] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 961f313b-b43f-4531-8a4b-0a39421d6a34/961f313b-b43f-4531-8a4b-0a39421d6a34.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 722.965798] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 722.966071] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25797b8d-5fff-4ba9-b907-f0cc37b8d47c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.973289] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 722.973289] env[62814]: value = "task-4293458" [ 722.973289] env[62814]: _type = "Task" [ 722.973289] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.981734] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293458, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.042564] env[62814]: DEBUG nova.compute.utils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 723.046744] env[62814]: DEBUG nova.compute.manager [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 723.047016] env[62814]: DEBUG nova.network.neutron [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 723.096492] env[62814]: DEBUG nova.policy [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8602e639f53943c49d8e7d5ac9fe20a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b9ca872e0c44873b87dc2a12edec042', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 723.115310] env[62814]: DEBUG oslo_concurrency.lockutils [req-4597df75-6f3f-4769-a127-b898b9d04961 req-2b5a95af-336e-4ddc-a35f-1dbaaeb6851f service nova] Releasing lock "refresh_cache-004fd137-4902-4313-a6f7-6c83cd76743d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 723.275494] env[62814]: DEBUG oslo_vmware.api [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293457, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.37011} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.278574] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 723.278574] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 723.278574] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 723.278847] env[62814]: INFO nova.compute.manager [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Took 1.14 seconds to destroy the instance on the hypervisor. [ 723.279098] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 723.279893] env[62814]: DEBUG nova.compute.manager [-] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 723.279893] env[62814]: DEBUG nova.network.neutron [-] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 723.393101] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5244f567-7ee3-3da6-b526-4ffa9fb8bd7d, 'name': SearchDatastore_Task, 'duration_secs': 0.022203} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.393452] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 723.394939] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 723.395283] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.396258] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 723.396478] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 723.396762] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0fb6b3ed-4491-42c3-a10b-0fbe78d3202b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.404999] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 723.405202] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 723.405936] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ce47b5a-ff60-449d-a27b-228cdc239228 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.413871] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for the task: (returnval){ [ 723.413871] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5232ef2a-01f7-b530-818f-d1614a0ba35e" [ 723.413871] env[62814]: _type = "Task" [ 723.413871] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.422571] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5232ef2a-01f7-b530-818f-d1614a0ba35e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.426006] env[62814]: DEBUG nova.network.neutron [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Successfully created port: a8d91825-084d-4bc1-be73-5f49c229e92e {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.484654] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293458, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06693} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.485161] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 723.486013] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f0cd3d-1061-45a7-abf4-8285c7a155a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.511147] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Reconfiguring VM instance instance-0000001b to attach disk [datastore2] 961f313b-b43f-4531-8a4b-0a39421d6a34/961f313b-b43f-4531-8a4b-0a39421d6a34.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.514255] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-373a272e-f008-4610-a581-aeb311bc98cf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.539444] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 723.539444] env[62814]: value = "task-4293459" [ 723.539444] env[62814]: _type = "Task" [ 723.539444] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.547718] env[62814]: DEBUG nova.compute.manager [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 723.561172] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293459, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.648884] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8368261a-4f86-4c2f-b09e-7ea5b10aae89 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.659587] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e2f933-3940-48a3-a25c-c2f3357f2a87 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.713240] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490f1164-b15a-40f2-beea-d3e90c0c6a23 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.727378] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcd42b9-7997-40e3-9a4e-8fa8afee302b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.748024] env[62814]: DEBUG nova.compute.provider_tree [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.924438] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5232ef2a-01f7-b530-818f-d1614a0ba35e, 'name': SearchDatastore_Task, 'duration_secs': 0.009363} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.925305] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5709aedc-d2e2-43ff-a8de-c3b376985017 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.931771] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for the task: (returnval){ [ 723.931771] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525e0c8f-5650-d020-cfff-3396638ebee8" [ 723.931771] env[62814]: _type = "Task" [ 723.931771] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.939058] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525e0c8f-5650-d020-cfff-3396638ebee8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.049247] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293459, 'name': ReconfigVM_Task, 'duration_secs': 0.281895} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.049811] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Reconfigured VM instance instance-0000001b to attach disk [datastore2] 961f313b-b43f-4531-8a4b-0a39421d6a34/961f313b-b43f-4531-8a4b-0a39421d6a34.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 724.050277] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4ae57cb-c7db-4121-a0eb-75d36127fa14 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.057854] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 724.057854] env[62814]: value = "task-4293460" [ 724.057854] env[62814]: _type = "Task" [ 724.057854] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.071159] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293460, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.096252] env[62814]: DEBUG nova.network.neutron [-] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.153746] env[62814]: DEBUG nova.compute.manager [req-9d2dbee2-9793-4b82-9dc1-9d1d5ccef41d req-320ed2a3-e400-425f-a176-ba74f87667bc service nova] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Received event network-vif-deleted-fb9b36b9-9a8a-4f9b-a8a3-d73a8dda47c4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 724.252268] env[62814]: DEBUG nova.scheduler.client.report [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 724.443045] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525e0c8f-5650-d020-cfff-3396638ebee8, 'name': SearchDatastore_Task, 'duration_secs': 0.010666} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.443249] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.443517] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 004fd137-4902-4313-a6f7-6c83cd76743d/004fd137-4902-4313-a6f7-6c83cd76743d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 724.443777] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67afc390-960b-47fc-aad2-00629fdca200 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.451722] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for the task: (returnval){ [ 724.451722] env[62814]: value = "task-4293461" [ 724.451722] env[62814]: _type = "Task" [ 724.451722] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.459677] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293461, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.567393] env[62814]: DEBUG nova.compute.manager [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 724.574822] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293460, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.593562] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 724.593840] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.594439] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 724.594439] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.594439] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 724.594439] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 724.594704] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 724.595443] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 724.595443] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 724.595443] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 724.595443] env[62814]: DEBUG nova.virt.hardware [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 724.596195] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca2015b-ebcd-424f-84d0-4137ea3ff128 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.599307] env[62814]: INFO nova.compute.manager [-] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Took 1.32 seconds to deallocate network for instance. [ 724.607376] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc37aad-d2f3-44e2-bfbb-bec61055a0f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.757644] env[62814]: DEBUG oslo_concurrency.lockutils [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.219s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 724.760206] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.381s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 724.760466] env[62814]: DEBUG nova.objects.instance [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lazy-loading 'resources' on Instance uuid 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 724.782410] env[62814]: INFO nova.scheduler.client.report [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleted allocations for instance a36ad785-2f33-4dbc-bc82-ab4a35020b0f [ 724.962970] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293461, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470315} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.963098] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 004fd137-4902-4313-a6f7-6c83cd76743d/004fd137-4902-4313-a6f7-6c83cd76743d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 724.963957] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 724.963957] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a1587f7-e5d3-4570-8542-a2579c0df8fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.970629] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for the task: (returnval){ [ 724.970629] env[62814]: value = "task-4293462" [ 724.970629] env[62814]: _type = "Task" [ 724.970629] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.981228] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293462, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.071804] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293460, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.096672] env[62814]: DEBUG nova.compute.manager [req-bbd37c33-8cb7-4df4-8608-8c4e4b7b6c99 req-1eda3b07-50d1-437a-8c70-4415604eb99c service nova] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Received event network-vif-plugged-a8d91825-084d-4bc1-be73-5f49c229e92e {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 725.097243] env[62814]: DEBUG oslo_concurrency.lockutils [req-bbd37c33-8cb7-4df4-8608-8c4e4b7b6c99 req-1eda3b07-50d1-437a-8c70-4415604eb99c service nova] Acquiring lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.097404] env[62814]: DEBUG oslo_concurrency.lockutils [req-bbd37c33-8cb7-4df4-8608-8c4e4b7b6c99 req-1eda3b07-50d1-437a-8c70-4415604eb99c service nova] Lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 725.097564] env[62814]: DEBUG oslo_concurrency.lockutils [req-bbd37c33-8cb7-4df4-8608-8c4e4b7b6c99 req-1eda3b07-50d1-437a-8c70-4415604eb99c service nova] Lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.097738] env[62814]: DEBUG nova.compute.manager [req-bbd37c33-8cb7-4df4-8608-8c4e4b7b6c99 req-1eda3b07-50d1-437a-8c70-4415604eb99c service nova] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] No waiting events found dispatching network-vif-plugged-a8d91825-084d-4bc1-be73-5f49c229e92e {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 725.097908] env[62814]: WARNING nova.compute.manager [req-bbd37c33-8cb7-4df4-8608-8c4e4b7b6c99 req-1eda3b07-50d1-437a-8c70-4415604eb99c service nova] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Received unexpected event network-vif-plugged-a8d91825-084d-4bc1-be73-5f49c229e92e for instance with vm_state building and task_state spawning. [ 725.109517] env[62814]: DEBUG oslo_concurrency.lockutils [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 725.146607] env[62814]: DEBUG nova.network.neutron [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Successfully updated port: a8d91825-084d-4bc1-be73-5f49c229e92e {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 725.292943] env[62814]: DEBUG oslo_concurrency.lockutils [None req-65d48ee7-60fb-42c0-a46f-9e3dd3da6282 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "a36ad785-2f33-4dbc-bc82-ab4a35020b0f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.272s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 725.481085] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293462, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06802} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.483390] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 725.484486] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e328f8f1-e56d-47e7-85e2-10506965588a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.505911] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Reconfiguring VM instance instance-0000001c to attach disk [datastore2] 004fd137-4902-4313-a6f7-6c83cd76743d/004fd137-4902-4313-a6f7-6c83cd76743d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 725.508744] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0983a08e-8d0a-4cb7-af70-cb7c0d9c515f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.528022] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for the task: (returnval){ [ 725.528022] env[62814]: value = "task-4293463" [ 725.528022] env[62814]: _type = "Task" [ 725.528022] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.537867] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293463, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.573896] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293460, 'name': Rename_Task, 'duration_secs': 1.139582} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.574159] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 725.574408] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52a586b5-12ab-4b23-ab34-c4140c552e0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.580938] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 725.580938] env[62814]: value = "task-4293464" [ 725.580938] env[62814]: _type = "Task" [ 725.580938] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.588471] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293464, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.649631] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquiring lock "refresh_cache-c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.649801] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquired lock "refresh_cache-c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.649993] env[62814]: DEBUG nova.network.neutron [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.719443] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1579033d-e112-47c4-bdcf-b3e85a83e95e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.728281] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f2d823-7e92-4b94-a513-0487d65b77e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.760394] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c320384a-d6de-4fe8-8dab-03d8e633cf43 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.768127] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b245f070-a724-427d-b05f-e49608c51056 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.783709] env[62814]: DEBUG nova.compute.provider_tree [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 726.038637] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293463, 'name': ReconfigVM_Task, 'duration_secs': 0.326556} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.038939] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Reconfigured VM instance instance-0000001c to attach disk [datastore2] 004fd137-4902-4313-a6f7-6c83cd76743d/004fd137-4902-4313-a6f7-6c83cd76743d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 726.039652] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dedd98b3-99f1-4ead-913d-a59c1b6babb2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.046483] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for the task: (returnval){ [ 726.046483] env[62814]: value = "task-4293465" [ 726.046483] env[62814]: _type = "Task" [ 726.046483] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.054945] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293465, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.089849] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293464, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.198646] env[62814]: DEBUG nova.network.neutron [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.286878] env[62814]: DEBUG nova.scheduler.client.report [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 726.377566] env[62814]: DEBUG nova.network.neutron [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Updating instance_info_cache with network_info: [{"id": "a8d91825-084d-4bc1-be73-5f49c229e92e", "address": "fa:16:3e:5d:8b:bd", "network": {"id": "4237f075-67ce-4796-a773-53de15be2135", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-888513471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9ca872e0c44873b87dc2a12edec042", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8d91825-08", "ovs_interfaceid": "a8d91825-084d-4bc1-be73-5f49c229e92e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.556231] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293465, 'name': Rename_Task, 'duration_secs': 0.144287} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.556533] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 726.556800] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0841671c-f932-4c54-ab42-d66ff5c46051 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.563149] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for the task: (returnval){ [ 726.563149] env[62814]: value = "task-4293466" [ 726.563149] env[62814]: _type = "Task" [ 726.563149] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.570552] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293466, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.589400] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293464, 'name': PowerOnVM_Task} progress is 1%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.794930] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.032s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.803296] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.022s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.803296] env[62814]: DEBUG nova.objects.instance [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lazy-loading 'resources' on Instance uuid 98d134b4-b4ca-4247-a638-ad5c24a694e5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 726.840648] env[62814]: INFO nova.scheduler.client.report [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleted allocations for instance 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e [ 726.880370] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Releasing lock "refresh_cache-c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 726.881036] env[62814]: DEBUG nova.compute.manager [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Instance network_info: |[{"id": "a8d91825-084d-4bc1-be73-5f49c229e92e", "address": "fa:16:3e:5d:8b:bd", "network": {"id": "4237f075-67ce-4796-a773-53de15be2135", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-888513471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9ca872e0c44873b87dc2a12edec042", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8d91825-08", "ovs_interfaceid": "a8d91825-084d-4bc1-be73-5f49c229e92e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 726.881658] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5d:8b:bd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4712af2-45ef-4652-8d2c-482ec70056d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8d91825-084d-4bc1-be73-5f49c229e92e', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 726.896114] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Creating folder: Project (0b9ca872e0c44873b87dc2a12edec042). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 726.897565] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-26a7e9d8-8a5f-47d2-b281-0496ecfe8b8f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.909636] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Created folder: Project (0b9ca872e0c44873b87dc2a12edec042) in parent group-v845547. [ 726.909826] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Creating folder: Instances. Parent ref: group-v845629. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 726.910142] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-73669a7a-f972-4955-b929-78a79fcc39f2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.919682] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Created folder: Instances in parent group-v845629. [ 726.919950] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 726.920170] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 726.920394] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15b7b0c6-5e6d-4875-8bf9-6d9430bb71cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.941539] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 726.941539] env[62814]: value = "task-4293469" [ 726.941539] env[62814]: _type = "Task" [ 726.941539] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.950203] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293469, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.073883] env[62814]: DEBUG oslo_vmware.api [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293466, 'name': PowerOnVM_Task, 'duration_secs': 0.483165} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.074273] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 727.074391] env[62814]: INFO nova.compute.manager [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Took 7.43 seconds to spawn the instance on the hypervisor. [ 727.074574] env[62814]: DEBUG nova.compute.manager [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 727.075533] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6773ee23-f1c6-4927-83e3-14ab8e21d0e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.097657] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293464, 'name': PowerOnVM_Task} progress is 64%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.211414] env[62814]: DEBUG nova.compute.manager [req-9a353f0d-7dbc-49f7-8fb2-931e195f7278 req-28e8464e-7b6b-497a-81ef-18f42034ce02 service nova] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Received event network-changed-a8d91825-084d-4bc1-be73-5f49c229e92e {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 727.214545] env[62814]: DEBUG nova.compute.manager [req-9a353f0d-7dbc-49f7-8fb2-931e195f7278 req-28e8464e-7b6b-497a-81ef-18f42034ce02 service nova] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Refreshing instance network info cache due to event network-changed-a8d91825-084d-4bc1-be73-5f49c229e92e. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 727.214955] env[62814]: DEBUG oslo_concurrency.lockutils [req-9a353f0d-7dbc-49f7-8fb2-931e195f7278 req-28e8464e-7b6b-497a-81ef-18f42034ce02 service nova] Acquiring lock "refresh_cache-c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.215214] env[62814]: DEBUG oslo_concurrency.lockutils [req-9a353f0d-7dbc-49f7-8fb2-931e195f7278 req-28e8464e-7b6b-497a-81ef-18f42034ce02 service nova] Acquired lock "refresh_cache-c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.215477] env[62814]: DEBUG nova.network.neutron [req-9a353f0d-7dbc-49f7-8fb2-931e195f7278 req-28e8464e-7b6b-497a-81ef-18f42034ce02 service nova] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Refreshing network info cache for port a8d91825-084d-4bc1-be73-5f49c229e92e {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 727.349796] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1dcbe8f-f446-472e-bc1e-861dd398c0a1 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.526s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.457707] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293469, 'name': CreateVM_Task, 'duration_secs': 0.404144} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.457877] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 727.458763] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.458931] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.459321] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 727.459579] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50357495-1232-4236-9bdd-16015fd183c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.466438] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for the task: (returnval){ [ 727.466438] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c7086d-2dad-c046-aba1-bbe86585b68b" [ 727.466438] env[62814]: _type = "Task" [ 727.466438] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.474531] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c7086d-2dad-c046-aba1-bbe86585b68b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.602280] env[62814]: DEBUG oslo_vmware.api [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293464, 'name': PowerOnVM_Task, 'duration_secs': 1.675284} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.605832] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 727.605832] env[62814]: INFO nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Took 10.78 seconds to spawn the instance on the hypervisor. [ 727.605832] env[62814]: DEBUG nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 727.609243] env[62814]: INFO nova.compute.manager [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Took 44.37 seconds to build instance. [ 727.609243] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d0232d-6a85-4e23-ac39-9be3bcb5470e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.888922] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33707579-8766-4925-9640-9680c4b0b9ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.898326] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ae709d-e70f-4eb6-9c72-8db979cd8f18 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.940396] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32da4fc5-589b-47ca-baad-5ae7c1c2e006 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.952341] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664b3250-8814-4c6e-9b41-e40a7719dfa0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.966394] env[62814]: DEBUG nova.compute.provider_tree [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.985467] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c7086d-2dad-c046-aba1-bbe86585b68b, 'name': SearchDatastore_Task, 'duration_secs': 0.011334} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.985782] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.986238] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 727.986304] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.986614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 727.986614] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 727.986867] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d04b1e39-1ba1-4a9c-9b28-5b4bb1baeae4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.996433] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 727.996669] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 727.998682] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1d09199-6161-4e6d-a9b1-debf5d609ab7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.004902] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for the task: (returnval){ [ 728.004902] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525a8717-772c-2e8c-5954-213c539c70dd" [ 728.004902] env[62814]: _type = "Task" [ 728.004902] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.014487] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525a8717-772c-2e8c-5954-213c539c70dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.046903] env[62814]: DEBUG nova.network.neutron [req-9a353f0d-7dbc-49f7-8fb2-931e195f7278 req-28e8464e-7b6b-497a-81ef-18f42034ce02 service nova] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Updated VIF entry in instance network info cache for port a8d91825-084d-4bc1-be73-5f49c229e92e. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 728.047463] env[62814]: DEBUG nova.network.neutron [req-9a353f0d-7dbc-49f7-8fb2-931e195f7278 req-28e8464e-7b6b-497a-81ef-18f42034ce02 service nova] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Updating instance_info_cache with network_info: [{"id": "a8d91825-084d-4bc1-be73-5f49c229e92e", "address": "fa:16:3e:5d:8b:bd", "network": {"id": "4237f075-67ce-4796-a773-53de15be2135", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-888513471-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b9ca872e0c44873b87dc2a12edec042", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4712af2-45ef-4652-8d2c-482ec70056d0", "external-id": "nsx-vlan-transportzone-826", "segmentation_id": 826, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8d91825-08", "ovs_interfaceid": "a8d91825-084d-4bc1-be73-5f49c229e92e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.112342] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ac80310-0f27-4e8d-8105-492aa9f98bec tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lock "004fd137-4902-4313-a6f7-6c83cd76743d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.089s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.125390] env[62814]: INFO nova.compute.manager [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Took 45.52 seconds to build instance. [ 728.399221] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "2d0b24c5-3593-4ef2-a637-d3590242ad79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.399464] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "2d0b24c5-3593-4ef2-a637-d3590242ad79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.443028] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "c18acec7-cf95-4cdf-aa49-32419d364534" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.443498] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "c18acec7-cf95-4cdf-aa49-32419d364534" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.477397] env[62814]: DEBUG nova.scheduler.client.report [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 728.480819] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquiring lock "004fd137-4902-4313-a6f7-6c83cd76743d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.480819] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lock "004fd137-4902-4313-a6f7-6c83cd76743d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.480963] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquiring lock "004fd137-4902-4313-a6f7-6c83cd76743d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 728.481232] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lock "004fd137-4902-4313-a6f7-6c83cd76743d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.481418] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lock "004fd137-4902-4313-a6f7-6c83cd76743d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.483320] env[62814]: INFO nova.compute.manager [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Terminating instance [ 728.516163] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525a8717-772c-2e8c-5954-213c539c70dd, 'name': SearchDatastore_Task, 'duration_secs': 0.00907} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.516971] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d1f1232-c1ff-465b-aeb9-60c1424feafe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.523719] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for the task: (returnval){ [ 728.523719] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52072215-3066-8e19-b7e6-e4162f43c209" [ 728.523719] env[62814]: _type = "Task" [ 728.523719] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.531354] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52072215-3066-8e19-b7e6-e4162f43c209, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.551239] env[62814]: DEBUG oslo_concurrency.lockutils [req-9a353f0d-7dbc-49f7-8fb2-931e195f7278 req-28e8464e-7b6b-497a-81ef-18f42034ce02 service nova] Releasing lock "refresh_cache-c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.616089] env[62814]: DEBUG nova.compute.manager [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 728.627304] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b5e83e82-170c-439f-9991-528f5f634a2b tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "961f313b-b43f-4531-8a4b-0a39421d6a34" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 63.460s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.981106] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.179s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.983593] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.600s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.985223] env[62814]: INFO nova.compute.claims [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.988505] env[62814]: DEBUG nova.compute.manager [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 728.988702] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.989811] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0382845f-b79a-4553-9ab9-4feed5600335 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.997447] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 728.997832] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a720f69d-3474-4c2b-a860-510fec633df3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.004480] env[62814]: DEBUG oslo_vmware.api [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for the task: (returnval){ [ 729.004480] env[62814]: value = "task-4293470" [ 729.004480] env[62814]: _type = "Task" [ 729.004480] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.005501] env[62814]: INFO nova.scheduler.client.report [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Deleted allocations for instance 98d134b4-b4ca-4247-a638-ad5c24a694e5 [ 729.017206] env[62814]: DEBUG oslo_vmware.api [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293470, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.034062] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52072215-3066-8e19-b7e6-e4162f43c209, 'name': SearchDatastore_Task, 'duration_secs': 0.009185} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.034897] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 729.035198] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c7829a8b-a9f7-40b1-958c-732e2eaa2b3c/c7829a8b-a9f7-40b1-958c-732e2eaa2b3c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 729.035470] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d66fd826-eecf-4bbc-8298-763a3b2e9512 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.038144] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.038376] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.038583] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.038787] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.039830] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.041188] env[62814]: INFO nova.compute.manager [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Terminating instance [ 729.049501] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for the task: (returnval){ [ 729.049501] env[62814]: value = "task-4293471" [ 729.049501] env[62814]: _type = "Task" [ 729.049501] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.059853] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293471, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.130077] env[62814]: DEBUG nova.compute.manager [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 729.135283] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.518649] env[62814]: DEBUG oslo_vmware.api [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293470, 'name': PowerOffVM_Task, 'duration_secs': 0.206567} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.519160] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d10b9280-2f8e-4f76-bb25-02201649580e tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "98d134b4-b4ca-4247-a638-ad5c24a694e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.814s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.520132] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 729.520343] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 729.520847] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1721c4a-f5f3-4aea-896c-ee259de8a7e1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.544555] env[62814]: DEBUG nova.compute.manager [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 729.544794] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 729.545744] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a16c821-3d58-4fc4-a837-a5c236d500b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.557435] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 729.557435] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-007c6287-f2e3-4a3c-b9d2-5dc5f013fe72 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.562892] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293471, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468301} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.563768] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c7829a8b-a9f7-40b1-958c-732e2eaa2b3c/c7829a8b-a9f7-40b1-958c-732e2eaa2b3c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 729.563993] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 729.564259] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6100c824-5e52-425d-b716-ef905add9038 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.568219] env[62814]: DEBUG oslo_vmware.api [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 729.568219] env[62814]: value = "task-4293473" [ 729.568219] env[62814]: _type = "Task" [ 729.568219] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.572586] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for the task: (returnval){ [ 729.572586] env[62814]: value = "task-4293474" [ 729.572586] env[62814]: _type = "Task" [ 729.572586] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.578750] env[62814]: DEBUG oslo_vmware.api [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293473, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.584336] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293474, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.587854] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 729.588099] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 729.588327] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Deleting the datastore file [datastore2] 004fd137-4902-4313-a6f7-6c83cd76743d {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 729.588623] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b32e19b6-70f2-4d4a-826a-638075077cb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.594868] env[62814]: DEBUG oslo_vmware.api [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for the task: (returnval){ [ 729.594868] env[62814]: value = "task-4293475" [ 729.594868] env[62814]: _type = "Task" [ 729.594868] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.603036] env[62814]: DEBUG oslo_vmware.api [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293475, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.649394] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.083259] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293474, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06798} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.088948] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 730.089408] env[62814]: DEBUG oslo_vmware.api [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293473, 'name': PowerOffVM_Task, 'duration_secs': 0.203044} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.090708] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6c35843-28ef-4fdb-a16f-0f252f123b0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.093194] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 730.093415] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 730.093708] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9303c8f-921d-4856-8250-ede07291028f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.116824] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Reconfiguring VM instance instance-0000001d to attach disk [datastore2] c7829a8b-a9f7-40b1-958c-732e2eaa2b3c/c7829a8b-a9f7-40b1-958c-732e2eaa2b3c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 730.120100] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10571d1b-04f7-4258-8268-c709e86a4564 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.139766] env[62814]: DEBUG oslo_vmware.api [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Task: {'id': task-4293475, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236003} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.140512] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 730.140747] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 730.140940] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 730.141186] env[62814]: INFO nova.compute.manager [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 730.141430] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 730.141634] env[62814]: DEBUG nova.compute.manager [-] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 730.141728] env[62814]: DEBUG nova.network.neutron [-] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.145015] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for the task: (returnval){ [ 730.145015] env[62814]: value = "task-4293477" [ 730.145015] env[62814]: _type = "Task" [ 730.145015] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.157150] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293477, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.161733] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 730.161923] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 730.162111] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Deleting the datastore file [datastore2] e6af4651-9f3a-4ce0-add8-06f1cfef255f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 730.164059] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c0d37df-43d5-47a0-8c3e-a31d77beb6c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.169724] env[62814]: DEBUG oslo_vmware.api [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 730.169724] env[62814]: value = "task-4293478" [ 730.169724] env[62814]: _type = "Task" [ 730.169724] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.180762] env[62814]: DEBUG oslo_vmware.api [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293478, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.502723] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "be02bc93-1278-4b3d-afa3-270c84585d1d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.502830] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "be02bc93-1278-4b3d-afa3-270c84585d1d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 730.503029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "be02bc93-1278-4b3d-afa3-270c84585d1d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.503231] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "be02bc93-1278-4b3d-afa3-270c84585d1d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 730.503404] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "be02bc93-1278-4b3d-afa3-270c84585d1d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 730.507927] env[62814]: INFO nova.compute.manager [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Terminating instance [ 730.512051] env[62814]: DEBUG nova.compute.manager [req-24934d43-d425-43b9-9e1e-5bc82ef34fc1 req-a791c145-5fb5-4b43-b0a7-86c55e32a32e service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Received event network-vif-deleted-b8697072-9a1f-46f2-9da3-c372d47e8385 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 730.512259] env[62814]: INFO nova.compute.manager [req-24934d43-d425-43b9-9e1e-5bc82ef34fc1 req-a791c145-5fb5-4b43-b0a7-86c55e32a32e service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Neutron deleted interface b8697072-9a1f-46f2-9da3-c372d47e8385; detaching it from the instance and deleting it from the info cache [ 730.512431] env[62814]: DEBUG nova.network.neutron [req-24934d43-d425-43b9-9e1e-5bc82ef34fc1 req-a791c145-5fb5-4b43-b0a7-86c55e32a32e service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.630140] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f4e75f-9f59-4deb-99d0-9125b485c194 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.638640] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98c0f26-63eb-4199-beed-2bebf5ad934f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.681554] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dca060c-bae9-404c-8744-f43101bccd1d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.693691] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293477, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.693957] env[62814]: DEBUG oslo_vmware.api [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293478, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146401} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.696152] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 730.696392] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 730.696582] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 730.696756] env[62814]: INFO nova.compute.manager [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Took 1.15 seconds to destroy the instance on the hypervisor. [ 730.697036] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 730.697298] env[62814]: DEBUG nova.compute.manager [-] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 730.697397] env[62814]: DEBUG nova.network.neutron [-] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 730.700125] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977c16e3-9dc2-4349-8e29-21d077607515 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.715026] env[62814]: DEBUG nova.compute.provider_tree [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.994610] env[62814]: DEBUG nova.network.neutron [-] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.015784] env[62814]: DEBUG nova.compute.manager [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 731.016016] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 731.016260] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c6f82f3-8e48-4539-96e5-8cbe031b1738 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.020011] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ea83d6-4606-4522-aa8e-4af02c393509 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.028574] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 731.030319] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b696ca72-6d2c-4460-b9ee-80207b3a6485 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.034826] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7676608a-1fff-48c2-9585-d689a746697b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.056698] env[62814]: DEBUG oslo_vmware.api [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 731.056698] env[62814]: value = "task-4293479" [ 731.056698] env[62814]: _type = "Task" [ 731.056698] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.071017] env[62814]: DEBUG nova.compute.manager [req-24934d43-d425-43b9-9e1e-5bc82ef34fc1 req-a791c145-5fb5-4b43-b0a7-86c55e32a32e service nova] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Detach interface failed, port_id=b8697072-9a1f-46f2-9da3-c372d47e8385, reason: Instance 004fd137-4902-4313-a6f7-6c83cd76743d could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 731.080459] env[62814]: DEBUG oslo_vmware.api [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293479, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.183897] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293477, 'name': ReconfigVM_Task, 'duration_secs': 0.671012} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.184304] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Reconfigured VM instance instance-0000001d to attach disk [datastore2] c7829a8b-a9f7-40b1-958c-732e2eaa2b3c/c7829a8b-a9f7-40b1-958c-732e2eaa2b3c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 731.185601] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef5a2ed0-1294-4608-956c-f0b78ca9c457 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.192785] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for the task: (returnval){ [ 731.192785] env[62814]: value = "task-4293480" [ 731.192785] env[62814]: _type = "Task" [ 731.192785] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.203437] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293480, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.220761] env[62814]: DEBUG nova.scheduler.client.report [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 731.496363] env[62814]: INFO nova.compute.manager [-] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Took 1.35 seconds to deallocate network for instance. [ 731.524246] env[62814]: DEBUG nova.network.neutron [-] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.579318] env[62814]: DEBUG oslo_vmware.api [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293479, 'name': PowerOffVM_Task, 'duration_secs': 0.39443} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.579633] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 731.579816] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 731.580120] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b69ae284-b424-4fdb-88aa-f2f84d76d47e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.667514] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 731.667807] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 731.668067] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Deleting the datastore file [datastore2] be02bc93-1278-4b3d-afa3-270c84585d1d {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 731.668399] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6d25108a-b700-4744-8556-a63e4d765857 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.675150] env[62814]: DEBUG oslo_vmware.api [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for the task: (returnval){ [ 731.675150] env[62814]: value = "task-4293482" [ 731.675150] env[62814]: _type = "Task" [ 731.675150] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.682583] env[62814]: DEBUG oslo_vmware.api [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.700991] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293480, 'name': Rename_Task, 'duration_secs': 0.138718} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.701264] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 731.701493] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e9e57ab2-1ef8-4cbc-81de-9b37ac10973a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.707495] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for the task: (returnval){ [ 731.707495] env[62814]: value = "task-4293483" [ 731.707495] env[62814]: _type = "Task" [ 731.707495] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.714733] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293483, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.728703] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.745s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.729189] env[62814]: DEBUG nova.compute.manager [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 731.732147] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 31.540s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 732.003553] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 732.027301] env[62814]: INFO nova.compute.manager [-] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Took 1.33 seconds to deallocate network for instance. [ 732.185375] env[62814]: DEBUG oslo_vmware.api [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Task: {'id': task-4293482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16796} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.185840] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 732.185840] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 732.186073] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.186159] env[62814]: INFO nova.compute.manager [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Took 1.17 seconds to destroy the instance on the hypervisor. [ 732.186402] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 732.186590] env[62814]: DEBUG nova.compute.manager [-] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 732.186675] env[62814]: DEBUG nova.network.neutron [-] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.221954] env[62814]: DEBUG oslo_vmware.api [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293483, 'name': PowerOnVM_Task, 'duration_secs': 0.457683} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.222306] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 732.222569] env[62814]: INFO nova.compute.manager [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Took 7.66 seconds to spawn the instance on the hypervisor. [ 732.222807] env[62814]: DEBUG nova.compute.manager [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 732.223848] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03823d50-a7a3-4a92-9cf3-2e6833aa64eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.235409] env[62814]: DEBUG nova.compute.utils [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 732.241045] env[62814]: DEBUG nova.compute.manager [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 732.241045] env[62814]: DEBUG nova.network.neutron [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 732.319379] env[62814]: DEBUG nova.policy [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51057ff83f92498bbb79387a1ec727b3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6a30b1a30e7d4a6e9b1aa6a92a8bac7d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 732.536630] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 732.561671] env[62814]: DEBUG nova.compute.manager [req-1d1245e4-0439-4db9-ad37-ca1452ab6dd7 req-f05a1116-70be-4789-85e5-5d870f85a790 service nova] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Received event network-vif-deleted-6be6bb54-d3f3-4068-9322-ccdb97c8470d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 732.562064] env[62814]: DEBUG nova.compute.manager [req-1d1245e4-0439-4db9-ad37-ca1452ab6dd7 req-f05a1116-70be-4789-85e5-5d870f85a790 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Received event network-vif-deleted-2710b5a0-6082-4d38-b71c-312e2c456a04 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 732.562377] env[62814]: INFO nova.compute.manager [req-1d1245e4-0439-4db9-ad37-ca1452ab6dd7 req-f05a1116-70be-4789-85e5-5d870f85a790 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Neutron deleted interface 2710b5a0-6082-4d38-b71c-312e2c456a04; detaching it from the instance and deleting it from the info cache [ 732.564256] env[62814]: DEBUG nova.network.neutron [req-1d1245e4-0439-4db9-ad37-ca1452ab6dd7 req-f05a1116-70be-4789-85e5-5d870f85a790 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.709054] env[62814]: DEBUG nova.network.neutron [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Successfully created port: e5baa36f-9802-41c1-add7-6020a8b74ce8 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.744082] env[62814]: DEBUG nova.compute.manager [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 732.748961] env[62814]: INFO nova.compute.manager [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Took 44.53 seconds to build instance. [ 732.806884] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e5e9d4f-c70f-4af5-8cf6-b5e1c8ab448b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.817290] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e902e43d-857d-4af8-ba6c-92fe99686667 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.857154] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b6309d-8ea3-4262-8d7a-01f9ae71dbac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.865602] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26b1a164-388a-4491-afb7-0e7099c0118d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.880817] env[62814]: DEBUG nova.compute.provider_tree [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.949269] env[62814]: DEBUG nova.network.neutron [-] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.069984] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-84ecbd6e-bc11-411a-bde1-53cf513c4e33 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.082514] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426d7adc-0428-4187-869b-92361f88bde0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.115442] env[62814]: DEBUG nova.compute.manager [req-1d1245e4-0439-4db9-ad37-ca1452ab6dd7 req-f05a1116-70be-4789-85e5-5d870f85a790 service nova] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Detach interface failed, port_id=2710b5a0-6082-4d38-b71c-312e2c456a04, reason: Instance be02bc93-1278-4b3d-afa3-270c84585d1d could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 733.179452] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquiring lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.253220] env[62814]: INFO nova.virt.block_device [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Booting with volume fa377baf-042f-48ba-8dc4-21baec93c237 at /dev/sda [ 733.255146] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ca36efd-41af-46b1-b660-93b75f8a8fea tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.221s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 733.258411] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.079s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.258411] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquiring lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.260428] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.260428] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 733.278111] env[62814]: INFO nova.compute.manager [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Terminating instance [ 733.323478] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0659b9a3-e101-44e1-9e42-f0345462eb64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.332947] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb6a8d1-1472-408f-b9ae-7fd41494691f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.365270] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de6efdc7-79ea-4a21-acf5-7aa619b01cc7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.373546] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41d2c45-6a54-439e-b63d-5cfad1def312 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.384470] env[62814]: DEBUG nova.scheduler.client.report [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 733.410107] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6133d1cf-e6e1-411a-9e98-8e7c81b2e37d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.417370] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30592e42-af96-4a3b-a091-308317416571 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.430810] env[62814]: DEBUG nova.virt.block_device [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Updating existing volume attachment record: ed893252-d471-4f79-abc2-e02dfd434766 {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 733.450629] env[62814]: INFO nova.compute.manager [-] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Took 1.26 seconds to deallocate network for instance. [ 733.777896] env[62814]: DEBUG nova.compute.manager [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 733.781668] env[62814]: DEBUG nova.compute.manager [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 733.781868] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.782731] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6aa4d5-05a0-4db7-9e9e-ed605e3bbcb3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.790597] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 733.790866] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1039ea0a-a312-45cf-872d-07f467d0e602 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.797283] env[62814]: DEBUG oslo_vmware.api [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for the task: (returnval){ [ 733.797283] env[62814]: value = "task-4293484" [ 733.797283] env[62814]: _type = "Task" [ 733.797283] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.806575] env[62814]: DEBUG oslo_vmware.api [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293484, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.957120] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.302336] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.306016] env[62814]: DEBUG oslo_vmware.api [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293484, 'name': PowerOffVM_Task, 'duration_secs': 0.16997} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.306816] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 734.306816] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 734.306816] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a4c45c12-826f-45fd-bda4-f10b9335647b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.366500] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 734.366718] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 734.366901] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Deleting the datastore file [datastore2] c7829a8b-a9f7-40b1-958c-732e2eaa2b3c {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 734.367229] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-566d29b6-93b1-4ef3-85ab-8b6f477e5a0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.373825] env[62814]: DEBUG oslo_vmware.api [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for the task: (returnval){ [ 734.373825] env[62814]: value = "task-4293486" [ 734.373825] env[62814]: _type = "Task" [ 734.373825] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.382409] env[62814]: DEBUG oslo_vmware.api [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293486, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.396365] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.663s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.398258] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 30.830s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.398899] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.398899] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 734.399057] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.830s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.400480] env[62814]: INFO nova.compute.claims [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 734.407721] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a65fcbcc-a064-4f8f-a4da-bb5f3198e1c7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.415014] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68469492-fd83-4636-ae53-d124edc162fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.434732] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29331c02-c7e2-4568-b036-336436a67d4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.444108] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50ed9bc-fdec-4580-81d5-dc23e0747bc2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.474684] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179227MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 734.474848] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.542064] env[62814]: DEBUG nova.compute.manager [req-8d1925ef-a7bc-4599-b0fe-8cc6834c9f35 req-4065637c-aa3e-4667-8e06-dc3bbfcb44e0 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Received event network-vif-plugged-e5baa36f-9802-41c1-add7-6020a8b74ce8 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 734.542064] env[62814]: DEBUG oslo_concurrency.lockutils [req-8d1925ef-a7bc-4599-b0fe-8cc6834c9f35 req-4065637c-aa3e-4667-8e06-dc3bbfcb44e0 service nova] Acquiring lock "bd0933ca-aab7-4dd4-a570-1a58a720f377-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.542064] env[62814]: DEBUG oslo_concurrency.lockutils [req-8d1925ef-a7bc-4599-b0fe-8cc6834c9f35 req-4065637c-aa3e-4667-8e06-dc3bbfcb44e0 service nova] Lock "bd0933ca-aab7-4dd4-a570-1a58a720f377-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.542064] env[62814]: DEBUG oslo_concurrency.lockutils [req-8d1925ef-a7bc-4599-b0fe-8cc6834c9f35 req-4065637c-aa3e-4667-8e06-dc3bbfcb44e0 service nova] Lock "bd0933ca-aab7-4dd4-a570-1a58a720f377-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.542064] env[62814]: DEBUG nova.compute.manager [req-8d1925ef-a7bc-4599-b0fe-8cc6834c9f35 req-4065637c-aa3e-4667-8e06-dc3bbfcb44e0 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] No waiting events found dispatching network-vif-plugged-e5baa36f-9802-41c1-add7-6020a8b74ce8 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 734.542415] env[62814]: WARNING nova.compute.manager [req-8d1925ef-a7bc-4599-b0fe-8cc6834c9f35 req-4065637c-aa3e-4667-8e06-dc3bbfcb44e0 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Received unexpected event network-vif-plugged-e5baa36f-9802-41c1-add7-6020a8b74ce8 for instance with vm_state building and task_state block_device_mapping. [ 734.623869] env[62814]: DEBUG nova.network.neutron [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Successfully updated port: e5baa36f-9802-41c1-add7-6020a8b74ce8 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 734.883299] env[62814]: DEBUG oslo_vmware.api [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Task: {'id': task-4293486, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134004} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.883299] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 734.883449] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 734.883544] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 734.883691] env[62814]: INFO nova.compute.manager [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 734.883934] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 734.884142] env[62814]: DEBUG nova.compute.manager [-] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 734.884259] env[62814]: DEBUG nova.network.neutron [-] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 734.958388] env[62814]: INFO nova.scheduler.client.report [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Deleted allocation for migration 126a427f-3696-4882-ac2f-5c3e4dbfe77f [ 735.126169] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Acquiring lock "refresh_cache-bd0933ca-aab7-4dd4-a570-1a58a720f377" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.126380] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Acquired lock "refresh_cache-bd0933ca-aab7-4dd4-a570-1a58a720f377" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 735.127097] env[62814]: DEBUG nova.network.neutron [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.467604] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3180aeee-ef4a-4794-91e3-7a9f4a7205f0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 38.368s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 735.555099] env[62814]: DEBUG nova.compute.manager [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 735.555099] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 735.555099] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.555397] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 735.555397] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.555494] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 735.555670] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 735.555809] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 735.555958] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 735.556160] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 735.556325] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 735.556522] env[62814]: DEBUG nova.virt.hardware [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 735.557447] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b139e8ec-5e02-473b-a2e5-1ea1e15a4972 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.567574] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5bee24a-8392-4cb3-8e33-1fb995bac1b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.673109] env[62814]: DEBUG nova.network.neutron [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 735.685416] env[62814]: DEBUG nova.network.neutron [-] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.866592] env[62814]: DEBUG nova.network.neutron [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Updating instance_info_cache with network_info: [{"id": "e5baa36f-9802-41c1-add7-6020a8b74ce8", "address": "fa:16:3e:c1:ae:99", "network": {"id": "02d057fd-7673-448b-b888-f528eaa99f0d", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-330959375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a30b1a30e7d4a6e9b1aa6a92a8bac7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "778b9a40-d603-4765-ac88-bd6d42c457a2", "external-id": "nsx-vlan-transportzone-114", "segmentation_id": 114, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5baa36f-98", "ovs_interfaceid": "e5baa36f-9802-41c1-add7-6020a8b74ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.898893] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3fb1aae-dce4-4949-b847-897ac6ec7c3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.907941] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7db9ef73-abe3-4818-a01a-6a133c47abc6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.937681] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165f7e23-11c6-4b8c-bed0-a43e4b248803 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.944564] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e81937-44f3-454b-803c-85fe64085a01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.957333] env[62814]: DEBUG nova.compute.provider_tree [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.189041] env[62814]: INFO nova.compute.manager [-] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Took 1.30 seconds to deallocate network for instance. [ 736.369616] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Releasing lock "refresh_cache-bd0933ca-aab7-4dd4-a570-1a58a720f377" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 736.370207] env[62814]: DEBUG nova.compute.manager [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Instance network_info: |[{"id": "e5baa36f-9802-41c1-add7-6020a8b74ce8", "address": "fa:16:3e:c1:ae:99", "network": {"id": "02d057fd-7673-448b-b888-f528eaa99f0d", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-330959375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a30b1a30e7d4a6e9b1aa6a92a8bac7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "778b9a40-d603-4765-ac88-bd6d42c457a2", "external-id": "nsx-vlan-transportzone-114", "segmentation_id": 114, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5baa36f-98", "ovs_interfaceid": "e5baa36f-9802-41c1-add7-6020a8b74ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 736.370408] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c1:ae:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '778b9a40-d603-4765-ac88-bd6d42c457a2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e5baa36f-9802-41c1-add7-6020a8b74ce8', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 736.379088] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Creating folder: Project (6a30b1a30e7d4a6e9b1aa6a92a8bac7d). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 736.379219] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-55c7f81e-cb10-479f-a2f7-5d50b4367cce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.394375] env[62814]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 736.394587] env[62814]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62814) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 736.395601] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Folder already exists: Project (6a30b1a30e7d4a6e9b1aa6a92a8bac7d). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 736.395872] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Creating folder: Instances. Parent ref: group-v845581. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 736.396238] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69de957c-0e6d-4f06-bb8d-7d2b8269bcbe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.407441] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Created folder: Instances in parent group-v845581. [ 736.407664] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 736.407834] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 736.408017] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-278c5bba-ce21-429d-b943-82611b85db71 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.425823] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 736.425823] env[62814]: value = "task-4293489" [ 736.425823] env[62814]: _type = "Task" [ 736.425823] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.433335] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293489, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.463809] env[62814]: DEBUG nova.scheduler.client.report [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 736.640530] env[62814]: DEBUG nova.compute.manager [req-039fc42e-93f0-4cec-8ce5-f003a2450c96 req-672d1e2c-e099-4a48-89d8-b09e90b133d7 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Received event network-changed-e5baa36f-9802-41c1-add7-6020a8b74ce8 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 736.640810] env[62814]: DEBUG nova.compute.manager [req-039fc42e-93f0-4cec-8ce5-f003a2450c96 req-672d1e2c-e099-4a48-89d8-b09e90b133d7 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Refreshing instance network info cache due to event network-changed-e5baa36f-9802-41c1-add7-6020a8b74ce8. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 736.641215] env[62814]: DEBUG oslo_concurrency.lockutils [req-039fc42e-93f0-4cec-8ce5-f003a2450c96 req-672d1e2c-e099-4a48-89d8-b09e90b133d7 service nova] Acquiring lock "refresh_cache-bd0933ca-aab7-4dd4-a570-1a58a720f377" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.641435] env[62814]: DEBUG oslo_concurrency.lockutils [req-039fc42e-93f0-4cec-8ce5-f003a2450c96 req-672d1e2c-e099-4a48-89d8-b09e90b133d7 service nova] Acquired lock "refresh_cache-bd0933ca-aab7-4dd4-a570-1a58a720f377" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 736.641890] env[62814]: DEBUG nova.network.neutron [req-039fc42e-93f0-4cec-8ce5-f003a2450c96 req-672d1e2c-e099-4a48-89d8-b09e90b133d7 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Refreshing network info cache for port e5baa36f-9802-41c1-add7-6020a8b74ce8 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 736.695693] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.936161] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293489, 'name': CreateVM_Task, 'duration_secs': 0.334248} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.936352] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 736.937086] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'ed893252-d471-4f79-abc2-e02dfd434766', 'mount_device': '/dev/sda', 'device_type': None, 'delete_on_termination': True, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845587', 'volume_id': 'fa377baf-042f-48ba-8dc4-21baec93c237', 'name': 'volume-fa377baf-042f-48ba-8dc4-21baec93c237', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bd0933ca-aab7-4dd4-a570-1a58a720f377', 'attached_at': '', 'detached_at': '', 'volume_id': 'fa377baf-042f-48ba-8dc4-21baec93c237', 'serial': 'fa377baf-042f-48ba-8dc4-21baec93c237'}, 'guest_format': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=62814) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 736.937315] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Root volume attach. Driver type: vmdk {{(pid=62814) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 736.938110] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3662fe4-801a-4b1d-9413-2ff4250f2a7f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.945728] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557b196d-dd40-4952-a208-c450fa897bf2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.951279] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dffc26e-2f27-45ab-9519-4f10a6e06b62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.956856] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-e031a9b2-00e8-409e-b82c-1d17f455a7e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.965245] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for the task: (returnval){ [ 736.965245] env[62814]: value = "task-4293490" [ 736.965245] env[62814]: _type = "Task" [ 736.965245] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.968463] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.569s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 736.968938] env[62814]: DEBUG nova.compute.manager [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 736.971471] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.249s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 736.972846] env[62814]: INFO nova.compute.claims [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.980427] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293490, 'name': RelocateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.151481] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 737.151771] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 737.409059] env[62814]: DEBUG nova.network.neutron [req-039fc42e-93f0-4cec-8ce5-f003a2450c96 req-672d1e2c-e099-4a48-89d8-b09e90b133d7 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Updated VIF entry in instance network info cache for port e5baa36f-9802-41c1-add7-6020a8b74ce8. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 737.409059] env[62814]: DEBUG nova.network.neutron [req-039fc42e-93f0-4cec-8ce5-f003a2450c96 req-672d1e2c-e099-4a48-89d8-b09e90b133d7 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Updating instance_info_cache with network_info: [{"id": "e5baa36f-9802-41c1-add7-6020a8b74ce8", "address": "fa:16:3e:c1:ae:99", "network": {"id": "02d057fd-7673-448b-b888-f528eaa99f0d", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-330959375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a30b1a30e7d4a6e9b1aa6a92a8bac7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "778b9a40-d603-4765-ac88-bd6d42c457a2", "external-id": "nsx-vlan-transportzone-114", "segmentation_id": 114, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5baa36f-98", "ovs_interfaceid": "e5baa36f-9802-41c1-add7-6020a8b74ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.475499] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293490, 'name': RelocateVM_Task} progress is 43%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.479867] env[62814]: DEBUG nova.compute.utils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 737.481339] env[62814]: DEBUG nova.compute.manager [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 737.481534] env[62814]: DEBUG nova.network.neutron [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 737.531901] env[62814]: DEBUG nova.policy [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da849afc04ce4ed490c00402c46fefb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a9b8185105e94d349ab02dad98962f91', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 737.911480] env[62814]: DEBUG oslo_concurrency.lockutils [req-039fc42e-93f0-4cec-8ce5-f003a2450c96 req-672d1e2c-e099-4a48-89d8-b09e90b133d7 service nova] Releasing lock "refresh_cache-bd0933ca-aab7-4dd4-a570-1a58a720f377" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.911756] env[62814]: DEBUG nova.compute.manager [req-039fc42e-93f0-4cec-8ce5-f003a2450c96 req-672d1e2c-e099-4a48-89d8-b09e90b133d7 service nova] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Received event network-vif-deleted-a8d91825-084d-4bc1-be73-5f49c229e92e {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 737.930823] env[62814]: DEBUG nova.network.neutron [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Successfully created port: 888abe76-da04-4e32-9f81-159cde73dbee {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.980872] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293490, 'name': RelocateVM_Task} progress is 56%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.984762] env[62814]: DEBUG nova.compute.manager [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 738.479553] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293490, 'name': RelocateVM_Task} progress is 71%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.551823] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fbaec1-5018-4e8a-a2c9-e6928d7b546f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.561897] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdfc378e-f05e-413b-bc83-6f448c32972c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.594736] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd34565-3d57-41d3-9790-10b14672d5cc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.603066] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1aa3c17-a7ed-4d2a-a439-d6dae10e14c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.617974] env[62814]: DEBUG nova.compute.provider_tree [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.977075] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293490, 'name': RelocateVM_Task} progress is 84%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.994631] env[62814]: DEBUG nova.compute.manager [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 739.023058] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 739.023344] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.023504] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 739.023685] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.023834] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 739.023981] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 739.024214] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 739.024374] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 739.024540] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 739.024919] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 739.025179] env[62814]: DEBUG nova.virt.hardware [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 739.026146] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd75815-8411-4975-9afd-a66e0982dcbc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.035398] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8772996-222e-4865-8b42-bac07af52954 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.121310] env[62814]: DEBUG nova.scheduler.client.report [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 739.484940] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293490, 'name': RelocateVM_Task} progress is 97%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.549606] env[62814]: DEBUG nova.compute.manager [req-57b29e9d-4210-4190-88f9-8edb0eff96b8 req-13a74e2d-8023-483c-80e0-0c7503b31e66 service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Received event network-vif-plugged-888abe76-da04-4e32-9f81-159cde73dbee {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 739.549828] env[62814]: DEBUG oslo_concurrency.lockutils [req-57b29e9d-4210-4190-88f9-8edb0eff96b8 req-13a74e2d-8023-483c-80e0-0c7503b31e66 service nova] Acquiring lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 739.550052] env[62814]: DEBUG oslo_concurrency.lockutils [req-57b29e9d-4210-4190-88f9-8edb0eff96b8 req-13a74e2d-8023-483c-80e0-0c7503b31e66 service nova] Lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 739.550226] env[62814]: DEBUG oslo_concurrency.lockutils [req-57b29e9d-4210-4190-88f9-8edb0eff96b8 req-13a74e2d-8023-483c-80e0-0c7503b31e66 service nova] Lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 739.550397] env[62814]: DEBUG nova.compute.manager [req-57b29e9d-4210-4190-88f9-8edb0eff96b8 req-13a74e2d-8023-483c-80e0-0c7503b31e66 service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] No waiting events found dispatching network-vif-plugged-888abe76-da04-4e32-9f81-159cde73dbee {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 739.550558] env[62814]: WARNING nova.compute.manager [req-57b29e9d-4210-4190-88f9-8edb0eff96b8 req-13a74e2d-8023-483c-80e0-0c7503b31e66 service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Received unexpected event network-vif-plugged-888abe76-da04-4e32-9f81-159cde73dbee for instance with vm_state building and task_state spawning. [ 739.627112] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.656s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 739.627712] env[62814]: DEBUG nova.compute.manager [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 739.630604] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.489s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 739.631989] env[62814]: INFO nova.compute.claims [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.664456] env[62814]: DEBUG nova.network.neutron [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Successfully updated port: 888abe76-da04-4e32-9f81-159cde73dbee {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 739.979744] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293490, 'name': RelocateVM_Task} progress is 97%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.136368] env[62814]: DEBUG nova.compute.utils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 740.139897] env[62814]: DEBUG nova.compute.manager [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 740.140099] env[62814]: DEBUG nova.network.neutron [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 740.166561] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquiring lock "refresh_cache-a589a3d8-20dc-4ff5-a192-c540e29f39d6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.166861] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquired lock "refresh_cache-a589a3d8-20dc-4ff5-a192-c540e29f39d6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 740.166955] env[62814]: DEBUG nova.network.neutron [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.191859] env[62814]: DEBUG nova.policy [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '40bcaa69e4234f62babef6c082688051', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f85eefae1aad42d8bed6bed203b8d221', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 740.480908] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293490, 'name': RelocateVM_Task} progress is 98%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.499863] env[62814]: DEBUG nova.network.neutron [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Successfully created port: efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 740.641357] env[62814]: DEBUG nova.compute.manager [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 740.741153] env[62814]: DEBUG nova.network.neutron [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.968608] env[62814]: DEBUG nova.network.neutron [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Updating instance_info_cache with network_info: [{"id": "888abe76-da04-4e32-9f81-159cde73dbee", "address": "fa:16:3e:18:b3:0f", "network": {"id": "e9a8dfb5-f43f-49ab-85a3-d619632af678", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-933191935-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9b8185105e94d349ab02dad98962f91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap888abe76-da", "ovs_interfaceid": "888abe76-da04-4e32-9f81-159cde73dbee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.980329] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293490, 'name': RelocateVM_Task, 'duration_secs': 3.586828} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.982812] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 740.983042] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845587', 'volume_id': 'fa377baf-042f-48ba-8dc4-21baec93c237', 'name': 'volume-fa377baf-042f-48ba-8dc4-21baec93c237', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bd0933ca-aab7-4dd4-a570-1a58a720f377', 'attached_at': '', 'detached_at': '', 'volume_id': 'fa377baf-042f-48ba-8dc4-21baec93c237', 'serial': 'fa377baf-042f-48ba-8dc4-21baec93c237'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 740.984590] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3e026e8-a826-4c0a-9302-dbe0816c434e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.004341] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16402e9-b817-431c-bc21-358a58c6dffb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.030377] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] volume-fa377baf-042f-48ba-8dc4-21baec93c237/volume-fa377baf-042f-48ba-8dc4-21baec93c237.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 741.033691] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55aac41c-6308-4e27-a62a-150f7b64e328 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.057367] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for the task: (returnval){ [ 741.057367] env[62814]: value = "task-4293491" [ 741.057367] env[62814]: _type = "Task" [ 741.057367] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.069233] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293491, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.190982] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9dec39-912d-4597-8233-2be1759d930c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.199848] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12fc9e8-d193-4d20-a265-cf1204ae5a2d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.232537] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58f4910a-aeab-46b2-bf87-a6bf43ca555e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.240508] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9bd8fe-3172-45cd-9d8a-ec880529944c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.255217] env[62814]: DEBUG nova.compute.provider_tree [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.472223] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Releasing lock "refresh_cache-a589a3d8-20dc-4ff5-a192-c540e29f39d6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 741.472860] env[62814]: DEBUG nova.compute.manager [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Instance network_info: |[{"id": "888abe76-da04-4e32-9f81-159cde73dbee", "address": "fa:16:3e:18:b3:0f", "network": {"id": "e9a8dfb5-f43f-49ab-85a3-d619632af678", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-933191935-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9b8185105e94d349ab02dad98962f91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap888abe76-da", "ovs_interfaceid": "888abe76-da04-4e32-9f81-159cde73dbee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 741.472997] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:18:b3:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '888abe76-da04-4e32-9f81-159cde73dbee', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 741.480663] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Creating folder: Project (a9b8185105e94d349ab02dad98962f91). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 741.480934] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a0a47451-86e2-4e47-8e01-607b422cd827 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.492626] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Created folder: Project (a9b8185105e94d349ab02dad98962f91) in parent group-v845547. [ 741.492801] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Creating folder: Instances. Parent ref: group-v845634. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 741.493036] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd2729ae-25f3-4546-be89-7e8f6118af59 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.504707] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Created folder: Instances in parent group-v845634. [ 741.504950] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 741.505255] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 741.505342] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5d394ae-4723-4378-9c18-961d1442999c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.525929] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 741.525929] env[62814]: value = "task-4293494" [ 741.525929] env[62814]: _type = "Task" [ 741.525929] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.533334] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293494, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.566128] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293491, 'name': ReconfigVM_Task, 'duration_secs': 0.361003} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.566416] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Reconfigured VM instance instance-0000001e to attach disk [datastore2] volume-fa377baf-042f-48ba-8dc4-21baec93c237/volume-fa377baf-042f-48ba-8dc4-21baec93c237.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 741.571202] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a3d719cd-929c-40b6-b767-d7b42c2e3826 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.583830] env[62814]: DEBUG nova.compute.manager [req-4b282d35-5962-4b31-a5da-8f599c3e4804 req-c9636cd7-6640-497b-bd68-810a8f09a170 service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Received event network-changed-888abe76-da04-4e32-9f81-159cde73dbee {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 741.584067] env[62814]: DEBUG nova.compute.manager [req-4b282d35-5962-4b31-a5da-8f599c3e4804 req-c9636cd7-6640-497b-bd68-810a8f09a170 service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Refreshing instance network info cache due to event network-changed-888abe76-da04-4e32-9f81-159cde73dbee. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 741.584305] env[62814]: DEBUG oslo_concurrency.lockutils [req-4b282d35-5962-4b31-a5da-8f599c3e4804 req-c9636cd7-6640-497b-bd68-810a8f09a170 service nova] Acquiring lock "refresh_cache-a589a3d8-20dc-4ff5-a192-c540e29f39d6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.584462] env[62814]: DEBUG oslo_concurrency.lockutils [req-4b282d35-5962-4b31-a5da-8f599c3e4804 req-c9636cd7-6640-497b-bd68-810a8f09a170 service nova] Acquired lock "refresh_cache-a589a3d8-20dc-4ff5-a192-c540e29f39d6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 741.584624] env[62814]: DEBUG nova.network.neutron [req-4b282d35-5962-4b31-a5da-8f599c3e4804 req-c9636cd7-6640-497b-bd68-810a8f09a170 service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Refreshing network info cache for port 888abe76-da04-4e32-9f81-159cde73dbee {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 741.591663] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for the task: (returnval){ [ 741.591663] env[62814]: value = "task-4293495" [ 741.591663] env[62814]: _type = "Task" [ 741.591663] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.603371] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293495, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.654330] env[62814]: DEBUG nova.compute.manager [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 741.681380] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 741.681692] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.681930] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 741.682224] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.682598] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 741.682824] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 741.683144] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 741.683340] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 741.683562] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 741.683812] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 741.684105] env[62814]: DEBUG nova.virt.hardware [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 741.685307] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a3cfbe-adcb-42b5-bfb2-b8565049a02d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.696146] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1d6633-353b-4d6f-b838-d6451865d65e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.760393] env[62814]: DEBUG nova.scheduler.client.report [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 742.036458] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293494, 'name': CreateVM_Task, 'duration_secs': 0.364353} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.036730] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 742.037584] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.037759] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.038083] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 742.038492] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46d95332-4764-45e0-9517-ec4c1c7a76d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.043491] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for the task: (returnval){ [ 742.043491] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5246c841-6880-13c7-caea-c6e9da28179a" [ 742.043491] env[62814]: _type = "Task" [ 742.043491] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.051259] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5246c841-6880-13c7-caea-c6e9da28179a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.101921] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293495, 'name': ReconfigVM_Task, 'duration_secs': 0.129998} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.104018] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845587', 'volume_id': 'fa377baf-042f-48ba-8dc4-21baec93c237', 'name': 'volume-fa377baf-042f-48ba-8dc4-21baec93c237', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bd0933ca-aab7-4dd4-a570-1a58a720f377', 'attached_at': '', 'detached_at': '', 'volume_id': 'fa377baf-042f-48ba-8dc4-21baec93c237', 'serial': 'fa377baf-042f-48ba-8dc4-21baec93c237'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 742.104018] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e73a12d1-febe-4d27-a661-043e45b32e2b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.110408] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for the task: (returnval){ [ 742.110408] env[62814]: value = "task-4293496" [ 742.110408] env[62814]: _type = "Task" [ 742.110408] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.120104] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293496, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.150227] env[62814]: DEBUG nova.network.neutron [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Successfully updated port: efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 742.266136] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.635s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.266708] env[62814]: DEBUG nova.compute.manager [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 742.276796] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.361s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.276796] env[62814]: INFO nova.compute.claims [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.369096] env[62814]: DEBUG nova.network.neutron [req-4b282d35-5962-4b31-a5da-8f599c3e4804 req-c9636cd7-6640-497b-bd68-810a8f09a170 service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Updated VIF entry in instance network info cache for port 888abe76-da04-4e32-9f81-159cde73dbee. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 742.369569] env[62814]: DEBUG nova.network.neutron [req-4b282d35-5962-4b31-a5da-8f599c3e4804 req-c9636cd7-6640-497b-bd68-810a8f09a170 service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Updating instance_info_cache with network_info: [{"id": "888abe76-da04-4e32-9f81-159cde73dbee", "address": "fa:16:3e:18:b3:0f", "network": {"id": "e9a8dfb5-f43f-49ab-85a3-d619632af678", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-933191935-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a9b8185105e94d349ab02dad98962f91", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap888abe76-da", "ovs_interfaceid": "888abe76-da04-4e32-9f81-159cde73dbee", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.557996] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5246c841-6880-13c7-caea-c6e9da28179a, 'name': SearchDatastore_Task, 'duration_secs': 0.040281} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.558916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.558916] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 742.559033] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.559231] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.559431] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 742.559691] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edd42bf8-2c85-4015-8401-27142ac1528d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.568535] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 742.568725] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 742.569489] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb37da20-33ae-4a45-b2e1-2d6e2bb0bd75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.574730] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for the task: (returnval){ [ 742.574730] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c29a17-aeef-d88f-8fee-d365c21992c3" [ 742.574730] env[62814]: _type = "Task" [ 742.574730] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.583791] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c29a17-aeef-d88f-8fee-d365c21992c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.619745] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293496, 'name': Rename_Task, 'duration_secs': 0.122843} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.621528] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 742.621806] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-160643e9-4e74-4fd2-9594-735db8452e34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.628657] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for the task: (returnval){ [ 742.628657] env[62814]: value = "task-4293497" [ 742.628657] env[62814]: _type = "Task" [ 742.628657] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.635830] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293497, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.652851] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquiring lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.653506] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquired lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.653506] env[62814]: DEBUG nova.network.neutron [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.772969] env[62814]: DEBUG nova.compute.utils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 742.774444] env[62814]: DEBUG nova.compute.manager [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 742.774546] env[62814]: DEBUG nova.network.neutron [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 742.818116] env[62814]: DEBUG nova.policy [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '58359539fd4545cdbf200ec364e43834', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dedbeff832a4ac48b0aa01bd6acc3f1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 742.872804] env[62814]: DEBUG oslo_concurrency.lockutils [req-4b282d35-5962-4b31-a5da-8f599c3e4804 req-c9636cd7-6640-497b-bd68-810a8f09a170 service nova] Releasing lock "refresh_cache-a589a3d8-20dc-4ff5-a192-c540e29f39d6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.085758] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c29a17-aeef-d88f-8fee-d365c21992c3, 'name': SearchDatastore_Task, 'duration_secs': 0.008657} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.086391] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-660fd7b8-545f-4514-9863-38acfea6763b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.091745] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for the task: (returnval){ [ 743.091745] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521b09ca-68a8-f700-7016-9664dc1785eb" [ 743.091745] env[62814]: _type = "Task" [ 743.091745] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.100208] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521b09ca-68a8-f700-7016-9664dc1785eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.141440] env[62814]: DEBUG oslo_vmware.api [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293497, 'name': PowerOnVM_Task, 'duration_secs': 0.416786} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.141866] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 743.141920] env[62814]: INFO nova.compute.manager [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Took 7.59 seconds to spawn the instance on the hypervisor. [ 743.142108] env[62814]: DEBUG nova.compute.manager [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 743.144403] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027d3e68-443e-4407-bee5-74435a755ad9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.190227] env[62814]: DEBUG nova.network.neutron [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.205021] env[62814]: DEBUG nova.network.neutron [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Successfully created port: 51cf8a1d-3070-4968-bd29-37c3fa579184 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 743.277873] env[62814]: DEBUG nova.compute.manager [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 743.399422] env[62814]: DEBUG nova.network.neutron [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Updating instance_info_cache with network_info: [{"id": "efe64bba-a178-4079-b256-51b9e74293c3", "address": "fa:16:3e:fe:f6:d8", "network": {"id": "755078a7-2f0f-43e8-b6a3-ca1a183cad9a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-249930933-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85eefae1aad42d8bed6bed203b8d221", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a78d5760-0bb1-4476-9578-8ad3c3144439", "external-id": "nsx-vlan-transportzone-325", "segmentation_id": 325, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapefe64bba-a1", "ovs_interfaceid": "efe64bba-a178-4079-b256-51b9e74293c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.603376] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521b09ca-68a8-f700-7016-9664dc1785eb, 'name': SearchDatastore_Task, 'duration_secs': 0.010098} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.606040] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.606323] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a589a3d8-20dc-4ff5-a192-c540e29f39d6/a589a3d8-20dc-4ff5-a192-c540e29f39d6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 743.607418] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17bf8ba0-6a35-489e-8021-228015a01079 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.613951] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for the task: (returnval){ [ 743.613951] env[62814]: value = "task-4293498" [ 743.613951] env[62814]: _type = "Task" [ 743.613951] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.625113] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293498, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.636359] env[62814]: DEBUG nova.compute.manager [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Received event network-vif-plugged-efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 743.637047] env[62814]: DEBUG oslo_concurrency.lockutils [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] Acquiring lock "369cd937-4c18-4068-ae59-70a1d585094b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.637047] env[62814]: DEBUG oslo_concurrency.lockutils [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] Lock "369cd937-4c18-4068-ae59-70a1d585094b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.637047] env[62814]: DEBUG oslo_concurrency.lockutils [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] Lock "369cd937-4c18-4068-ae59-70a1d585094b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.637355] env[62814]: DEBUG nova.compute.manager [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] No waiting events found dispatching network-vif-plugged-efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 743.637355] env[62814]: WARNING nova.compute.manager [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Received unexpected event network-vif-plugged-efe64bba-a178-4079-b256-51b9e74293c3 for instance with vm_state building and task_state spawning. [ 743.637505] env[62814]: DEBUG nova.compute.manager [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Received event network-changed-efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 743.637708] env[62814]: DEBUG nova.compute.manager [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Refreshing instance network info cache due to event network-changed-efe64bba-a178-4079-b256-51b9e74293c3. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 743.637892] env[62814]: DEBUG oslo_concurrency.lockutils [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] Acquiring lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.671617] env[62814]: INFO nova.compute.manager [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Took 46.30 seconds to build instance. [ 743.850890] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c56050-5ac4-41a9-8bd3-59dff2823f06 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.861229] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d29d72-c02f-40e5-b91a-d678433d3d8f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.895643] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b210eabf-3502-442d-b839-4673c476601e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.905581] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eafb617-a110-4174-a026-6d423b926a51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.911042] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Releasing lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 743.911418] env[62814]: DEBUG nova.compute.manager [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Instance network_info: |[{"id": "efe64bba-a178-4079-b256-51b9e74293c3", "address": "fa:16:3e:fe:f6:d8", "network": {"id": "755078a7-2f0f-43e8-b6a3-ca1a183cad9a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-249930933-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85eefae1aad42d8bed6bed203b8d221", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a78d5760-0bb1-4476-9578-8ad3c3144439", "external-id": "nsx-vlan-transportzone-325", "segmentation_id": 325, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapefe64bba-a1", "ovs_interfaceid": "efe64bba-a178-4079-b256-51b9e74293c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 743.911802] env[62814]: DEBUG oslo_concurrency.lockutils [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] Acquired lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.912014] env[62814]: DEBUG nova.network.neutron [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Refreshing network info cache for port efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 743.913352] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:f6:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a78d5760-0bb1-4476-9578-8ad3c3144439', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'efe64bba-a178-4079-b256-51b9e74293c3', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 743.921933] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Creating folder: Project (f85eefae1aad42d8bed6bed203b8d221). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 743.926105] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-edf32a1c-815f-4b10-8bd6-fed8cef67136 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.939331] env[62814]: DEBUG nova.compute.provider_tree [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 743.951472] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Created folder: Project (f85eefae1aad42d8bed6bed203b8d221) in parent group-v845547. [ 743.951472] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Creating folder: Instances. Parent ref: group-v845637. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 743.952132] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1b6d9a75-5325-4ef8-bd4b-d97e2a48896b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.963520] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Created folder: Instances in parent group-v845637. [ 743.963773] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 743.964035] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 743.964258] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b45d33a0-ae96-4a1f-8b58-2d606dd4e682 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.987140] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 743.987140] env[62814]: value = "task-4293501" [ 743.987140] env[62814]: _type = "Task" [ 743.987140] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.000705] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293501, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.124884] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293498, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452854} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.125165] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a589a3d8-20dc-4ff5-a192-c540e29f39d6/a589a3d8-20dc-4ff5-a192-c540e29f39d6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 744.126270] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.126270] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0eebe35a-8e9d-4096-b8d6-6c3d3d61f1d1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.132931] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for the task: (returnval){ [ 744.132931] env[62814]: value = "task-4293502" [ 744.132931] env[62814]: _type = "Task" [ 744.132931] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.144149] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293502, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.175490] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7dba9361-ae45-45c4-b24b-9a717cc5b5c8 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lock "bd0933ca-aab7-4dd4-a570-1a58a720f377" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.211s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.291204] env[62814]: DEBUG nova.compute.manager [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 744.311748] env[62814]: DEBUG nova.network.neutron [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Updated VIF entry in instance network info cache for port efe64bba-a178-4079-b256-51b9e74293c3. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 744.312367] env[62814]: DEBUG nova.network.neutron [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Updating instance_info_cache with network_info: [{"id": "efe64bba-a178-4079-b256-51b9e74293c3", "address": "fa:16:3e:fe:f6:d8", "network": {"id": "755078a7-2f0f-43e8-b6a3-ca1a183cad9a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-249930933-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85eefae1aad42d8bed6bed203b8d221", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a78d5760-0bb1-4476-9578-8ad3c3144439", "external-id": "nsx-vlan-transportzone-325", "segmentation_id": 325, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapefe64bba-a1", "ovs_interfaceid": "efe64bba-a178-4079-b256-51b9e74293c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.327087] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 744.327582] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.327853] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 744.328118] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.328353] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 744.328614] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 744.329133] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 744.329408] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 744.330084] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 744.330481] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 744.330906] env[62814]: DEBUG nova.virt.hardware [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 744.333542] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ce4a5d-6e4c-4507-b4f3-3526a1cb96e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.341923] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0d61be-f71d-4bde-be63-9e7aacf50f2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.445021] env[62814]: DEBUG nova.scheduler.client.report [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 744.497452] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293501, 'name': CreateVM_Task, 'duration_secs': 0.405783} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.497624] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 744.498345] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.498567] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 744.498898] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 744.499162] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-59d3b02a-a3fe-42a2-b115-2797be5945c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.503941] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for the task: (returnval){ [ 744.503941] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52794e94-ccce-de93-d1e6-31c3a51e04bc" [ 744.503941] env[62814]: _type = "Task" [ 744.503941] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.512371] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52794e94-ccce-de93-d1e6-31c3a51e04bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.643916] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293502, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069282} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.645371] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 744.646419] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb752a48-d1c6-4473-8195-7a0429cd44da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.672641] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] a589a3d8-20dc-4ff5-a192-c540e29f39d6/a589a3d8-20dc-4ff5-a192-c540e29f39d6.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.673369] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-215c0c4d-d17c-43db-9807-9a95e142e030 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.689470] env[62814]: DEBUG nova.compute.manager [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 744.703543] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for the task: (returnval){ [ 744.703543] env[62814]: value = "task-4293503" [ 744.703543] env[62814]: _type = "Task" [ 744.703543] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.713031] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293503, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.821053] env[62814]: DEBUG oslo_concurrency.lockutils [req-1a206e86-3d32-4aa9-bdb7-8ab62b7a778d req-85e2f4ba-92ed-4b49-817c-51459b212e3f service nova] Releasing lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.864342] env[62814]: DEBUG nova.network.neutron [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Successfully updated port: 51cf8a1d-3070-4968-bd29-37c3fa579184 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 744.910238] env[62814]: DEBUG nova.compute.manager [req-e90cd2e6-a686-41d6-b37c-010d4da0adc9 req-6da7e0ae-14a0-495f-a083-02cbd329b487 service nova] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Received event network-vif-plugged-51cf8a1d-3070-4968-bd29-37c3fa579184 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 744.910447] env[62814]: DEBUG oslo_concurrency.lockutils [req-e90cd2e6-a686-41d6-b37c-010d4da0adc9 req-6da7e0ae-14a0-495f-a083-02cbd329b487 service nova] Acquiring lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 744.910741] env[62814]: DEBUG oslo_concurrency.lockutils [req-e90cd2e6-a686-41d6-b37c-010d4da0adc9 req-6da7e0ae-14a0-495f-a083-02cbd329b487 service nova] Lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.910858] env[62814]: DEBUG oslo_concurrency.lockutils [req-e90cd2e6-a686-41d6-b37c-010d4da0adc9 req-6da7e0ae-14a0-495f-a083-02cbd329b487 service nova] Lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.910987] env[62814]: DEBUG nova.compute.manager [req-e90cd2e6-a686-41d6-b37c-010d4da0adc9 req-6da7e0ae-14a0-495f-a083-02cbd329b487 service nova] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] No waiting events found dispatching network-vif-plugged-51cf8a1d-3070-4968-bd29-37c3fa579184 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 744.911195] env[62814]: WARNING nova.compute.manager [req-e90cd2e6-a686-41d6-b37c-010d4da0adc9 req-6da7e0ae-14a0-495f-a083-02cbd329b487 service nova] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Received unexpected event network-vif-plugged-51cf8a1d-3070-4968-bd29-37c3fa579184 for instance with vm_state building and task_state spawning. [ 744.953192] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.679s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 744.953743] env[62814]: DEBUG nova.compute.manager [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 744.962173] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.609s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 744.962173] env[62814]: INFO nova.compute.claims [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.014714] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52794e94-ccce-de93-d1e6-31c3a51e04bc, 'name': SearchDatastore_Task, 'duration_secs': 0.012842} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.015137] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 745.015373] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 745.016479] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.016674] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 745.016883] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 745.017450] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f7598cb-955b-4419-bf33-9b9820608b5f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.026220] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 745.026411] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 745.027137] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0998557b-0cf6-4ab6-9d25-bff3d2da3666 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.033227] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for the task: (returnval){ [ 745.033227] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272545f-4cbd-700f-daf6-aecb867d90c9" [ 745.033227] env[62814]: _type = "Task" [ 745.033227] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.041565] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272545f-4cbd-700f-daf6-aecb867d90c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.213272] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293503, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.216678] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 745.367606] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "refresh_cache-217e5812-95cc-4104-8d7b-82dfca2c0fcc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.368085] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "refresh_cache-217e5812-95cc-4104-8d7b-82dfca2c0fcc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 745.368085] env[62814]: DEBUG nova.network.neutron [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.460704] env[62814]: DEBUG nova.compute.utils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 745.462185] env[62814]: DEBUG nova.compute.manager [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 745.462387] env[62814]: DEBUG nova.network.neutron [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 745.527674] env[62814]: DEBUG nova.policy [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b663046f1c5483a9b12968b16ccd98b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd14adacd89bc460d8d1a7c7d2ecc5d7f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 745.546032] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272545f-4cbd-700f-daf6-aecb867d90c9, 'name': SearchDatastore_Task, 'duration_secs': 0.015706} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.546454] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e435ca5-e11b-455d-aa41-cc9f2fc30d20 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.552591] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for the task: (returnval){ [ 745.552591] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528a3e42-206a-cd69-db29-3138bccb70ed" [ 745.552591] env[62814]: _type = "Task" [ 745.552591] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.560548] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528a3e42-206a-cd69-db29-3138bccb70ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.664756] env[62814]: DEBUG nova.compute.manager [req-5244865b-f79b-4666-89bf-e151f9b8b99a req-547b346a-11dc-42e3-9a14-a8f0bd4b13ca service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Received event network-changed-e5baa36f-9802-41c1-add7-6020a8b74ce8 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 745.664756] env[62814]: DEBUG nova.compute.manager [req-5244865b-f79b-4666-89bf-e151f9b8b99a req-547b346a-11dc-42e3-9a14-a8f0bd4b13ca service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Refreshing instance network info cache due to event network-changed-e5baa36f-9802-41c1-add7-6020a8b74ce8. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 745.664756] env[62814]: DEBUG oslo_concurrency.lockutils [req-5244865b-f79b-4666-89bf-e151f9b8b99a req-547b346a-11dc-42e3-9a14-a8f0bd4b13ca service nova] Acquiring lock "refresh_cache-bd0933ca-aab7-4dd4-a570-1a58a720f377" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.664756] env[62814]: DEBUG oslo_concurrency.lockutils [req-5244865b-f79b-4666-89bf-e151f9b8b99a req-547b346a-11dc-42e3-9a14-a8f0bd4b13ca service nova] Acquired lock "refresh_cache-bd0933ca-aab7-4dd4-a570-1a58a720f377" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 745.665119] env[62814]: DEBUG nova.network.neutron [req-5244865b-f79b-4666-89bf-e151f9b8b99a req-547b346a-11dc-42e3-9a14-a8f0bd4b13ca service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Refreshing network info cache for port e5baa36f-9802-41c1-add7-6020a8b74ce8 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 745.716363] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293503, 'name': ReconfigVM_Task, 'duration_secs': 0.897333} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.716763] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Reconfigured VM instance instance-0000001f to attach disk [datastore2] a589a3d8-20dc-4ff5-a192-c540e29f39d6/a589a3d8-20dc-4ff5-a192-c540e29f39d6.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 745.717466] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b987b67-5dcb-4a3f-b24c-2ead22293b5c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.724878] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for the task: (returnval){ [ 745.724878] env[62814]: value = "task-4293504" [ 745.724878] env[62814]: _type = "Task" [ 745.724878] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.732456] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293504, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.840802] env[62814]: DEBUG nova.network.neutron [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Successfully created port: ce7e8021-3ebf-4e8c-a0c2-618fa44c6957 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.921359] env[62814]: DEBUG nova.network.neutron [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.966625] env[62814]: DEBUG nova.compute.manager [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 746.069047] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528a3e42-206a-cd69-db29-3138bccb70ed, 'name': SearchDatastore_Task, 'duration_secs': 0.012163} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.069337] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 746.069818] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 369cd937-4c18-4068-ae59-70a1d585094b/369cd937-4c18-4068-ae59-70a1d585094b.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 746.070135] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1687d37e-c598-41ae-b67c-8501d5aa6e3f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.081406] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for the task: (returnval){ [ 746.081406] env[62814]: value = "task-4293505" [ 746.081406] env[62814]: _type = "Task" [ 746.081406] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.092427] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293505, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.101159] env[62814]: DEBUG nova.network.neutron [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Updating instance_info_cache with network_info: [{"id": "51cf8a1d-3070-4968-bd29-37c3fa579184", "address": "fa:16:3e:7e:55:48", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51cf8a1d-30", "ovs_interfaceid": "51cf8a1d-3070-4968-bd29-37c3fa579184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.214076] env[62814]: DEBUG nova.network.neutron [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Successfully created port: fbfe5ed2-1472-41c4-a028-38e81d37b3fe {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.236017] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293504, 'name': Rename_Task, 'duration_secs': 0.179759} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.236322] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 746.236641] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c69a5e3a-98c8-4ce2-9765-46b185bb54b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.248553] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for the task: (returnval){ [ 746.248553] env[62814]: value = "task-4293506" [ 746.248553] env[62814]: _type = "Task" [ 746.248553] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.261105] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293506, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.552021] env[62814]: DEBUG nova.network.neutron [req-5244865b-f79b-4666-89bf-e151f9b8b99a req-547b346a-11dc-42e3-9a14-a8f0bd4b13ca service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Updated VIF entry in instance network info cache for port e5baa36f-9802-41c1-add7-6020a8b74ce8. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 746.552021] env[62814]: DEBUG nova.network.neutron [req-5244865b-f79b-4666-89bf-e151f9b8b99a req-547b346a-11dc-42e3-9a14-a8f0bd4b13ca service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Updating instance_info_cache with network_info: [{"id": "e5baa36f-9802-41c1-add7-6020a8b74ce8", "address": "fa:16:3e:c1:ae:99", "network": {"id": "02d057fd-7673-448b-b888-f528eaa99f0d", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-330959375-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6a30b1a30e7d4a6e9b1aa6a92a8bac7d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "778b9a40-d603-4765-ac88-bd6d42c457a2", "external-id": "nsx-vlan-transportzone-114", "segmentation_id": 114, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5baa36f-98", "ovs_interfaceid": "e5baa36f-9802-41c1-add7-6020a8b74ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.574476] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab34c5e-c6d9-4fb2-b087-e44176f1c709 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.589740] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eabf6a8b-350e-43d9-a309-43f3ce5b8d0b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.596272] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293505, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516012} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.597293] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 369cd937-4c18-4068-ae59-70a1d585094b/369cd937-4c18-4068-ae59-70a1d585094b.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 746.597567] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 746.598100] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-05aae8c4-9840-4176-a0af-f7019be72619 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.632245] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "refresh_cache-217e5812-95cc-4104-8d7b-82dfca2c0fcc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 746.632562] env[62814]: DEBUG nova.compute.manager [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Instance network_info: |[{"id": "51cf8a1d-3070-4968-bd29-37c3fa579184", "address": "fa:16:3e:7e:55:48", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51cf8a1d-30", "ovs_interfaceid": "51cf8a1d-3070-4968-bd29-37c3fa579184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 746.633885] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:55:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3db2ab9e-1244-4377-b05f-ab76003f2428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51cf8a1d-3070-4968-bd29-37c3fa579184', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 746.642785] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 746.643504] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69603452-9281-4a14-8ae2-ccd63222a00a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.647319] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 746.647950] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for the task: (returnval){ [ 746.647950] env[62814]: value = "task-4293507" [ 746.647950] env[62814]: _type = "Task" [ 746.647950] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.647950] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f9e5c232-1475-45c4-8b96-581c414cfeb8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.672507] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4907d87-b2ac-4520-9852-6b1d3f2f5f82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.681068] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 746.681068] env[62814]: value = "task-4293508" [ 746.681068] env[62814]: _type = "Task" [ 746.681068] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.681476] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293507, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.693971] env[62814]: DEBUG nova.compute.provider_tree [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.700449] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293508, 'name': CreateVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.760189] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293506, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.830942] env[62814]: DEBUG nova.network.neutron [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Successfully created port: 92958ac5-6305-4ce3-aa70-0c6080099c48 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.988223] env[62814]: DEBUG nova.compute.manager [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 747.017146] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 747.017458] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.017652] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 747.017875] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.018168] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 747.018365] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 747.018608] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 747.018801] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 747.018991] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 747.019193] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 747.019435] env[62814]: DEBUG nova.virt.hardware [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 747.020368] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b77a4fd-99be-4931-ad18-969c33527c9f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.028872] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3320418b-8bd9-439a-853b-a1d728b47f28 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.056902] env[62814]: DEBUG oslo_concurrency.lockutils [req-5244865b-f79b-4666-89bf-e151f9b8b99a req-547b346a-11dc-42e3-9a14-a8f0bd4b13ca service nova] Releasing lock "refresh_cache-bd0933ca-aab7-4dd4-a570-1a58a720f377" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.075797] env[62814]: DEBUG nova.compute.manager [req-3e105ebc-24b7-4426-aa22-dfbea7ea449c req-fc389d61-7308-4931-8e84-5531c209b10d service nova] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Received event network-changed-51cf8a1d-3070-4968-bd29-37c3fa579184 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 747.075990] env[62814]: DEBUG nova.compute.manager [req-3e105ebc-24b7-4426-aa22-dfbea7ea449c req-fc389d61-7308-4931-8e84-5531c209b10d service nova] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Refreshing instance network info cache due to event network-changed-51cf8a1d-3070-4968-bd29-37c3fa579184. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 747.076226] env[62814]: DEBUG oslo_concurrency.lockutils [req-3e105ebc-24b7-4426-aa22-dfbea7ea449c req-fc389d61-7308-4931-8e84-5531c209b10d service nova] Acquiring lock "refresh_cache-217e5812-95cc-4104-8d7b-82dfca2c0fcc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.076373] env[62814]: DEBUG oslo_concurrency.lockutils [req-3e105ebc-24b7-4426-aa22-dfbea7ea449c req-fc389d61-7308-4931-8e84-5531c209b10d service nova] Acquired lock "refresh_cache-217e5812-95cc-4104-8d7b-82dfca2c0fcc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.076534] env[62814]: DEBUG nova.network.neutron [req-3e105ebc-24b7-4426-aa22-dfbea7ea449c req-fc389d61-7308-4931-8e84-5531c209b10d service nova] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Refreshing network info cache for port 51cf8a1d-3070-4968-bd29-37c3fa579184 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 747.177945] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293507, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066831} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.177945] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 747.177945] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba8796f-581c-4694-a945-5a4016fad542 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.206643] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] 369cd937-4c18-4068-ae59-70a1d585094b/369cd937-4c18-4068-ae59-70a1d585094b.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 747.212678] env[62814]: DEBUG nova.scheduler.client.report [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 747.215844] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77c8f868-62c9-4fbd-aaed-1275a2f261d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.238291] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293508, 'name': CreateVM_Task, 'duration_secs': 0.351579} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.239331] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 747.239688] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for the task: (returnval){ [ 747.239688] env[62814]: value = "task-4293509" [ 747.239688] env[62814]: _type = "Task" [ 747.239688] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.240323] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.240487] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.240800] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 747.241103] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-658dd8eb-8b63-45d8-923b-7f75aa1e8c2f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.252294] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293509, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.252634] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 747.252634] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aa2d82-16dc-8de3-8e70-f2db488cac92" [ 747.252634] env[62814]: _type = "Task" [ 747.252634] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.263037] env[62814]: DEBUG oslo_vmware.api [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293506, 'name': PowerOnVM_Task, 'duration_secs': 0.740561} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.266362] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 747.266568] env[62814]: INFO nova.compute.manager [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Took 8.27 seconds to spawn the instance on the hypervisor. [ 747.266743] env[62814]: DEBUG nova.compute.manager [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 747.267017] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aa2d82-16dc-8de3-8e70-f2db488cac92, 'name': SearchDatastore_Task, 'duration_secs': 0.011844} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.267791] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc49fa46-c41f-4633-9def-12231bda8ef6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.270130] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 747.270354] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 747.270583] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.270727] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.270898] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 747.271141] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dab707d3-2d92-4fc3-ae3e-de04097c8a74 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.278819] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 747.278986] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 747.279865] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc4ec342-9cdf-4897-b3e1-95f49c464d38 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.284174] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 747.284174] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520981b4-9413-92ad-2de7-6c84cc3b8cb1" [ 747.284174] env[62814]: _type = "Task" [ 747.284174] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.291190] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520981b4-9413-92ad-2de7-6c84cc3b8cb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.732765] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.774s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.733301] env[62814]: DEBUG nova.compute.manager [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 747.736323] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.914s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.737934] env[62814]: INFO nova.compute.claims [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.754748] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293509, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.790457] env[62814]: INFO nova.compute.manager [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Took 39.25 seconds to build instance. [ 747.794242] env[62814]: DEBUG nova.network.neutron [req-3e105ebc-24b7-4426-aa22-dfbea7ea449c req-fc389d61-7308-4931-8e84-5531c209b10d service nova] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Updated VIF entry in instance network info cache for port 51cf8a1d-3070-4968-bd29-37c3fa579184. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 747.794242] env[62814]: DEBUG nova.network.neutron [req-3e105ebc-24b7-4426-aa22-dfbea7ea449c req-fc389d61-7308-4931-8e84-5531c209b10d service nova] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Updating instance_info_cache with network_info: [{"id": "51cf8a1d-3070-4968-bd29-37c3fa579184", "address": "fa:16:3e:7e:55:48", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51cf8a1d-30", "ovs_interfaceid": "51cf8a1d-3070-4968-bd29-37c3fa579184", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.799280] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520981b4-9413-92ad-2de7-6c84cc3b8cb1, 'name': SearchDatastore_Task, 'duration_secs': 0.008035} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.800331] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53f54ed8-6ebf-46ef-a2f8-8dec129a60b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.806872] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 747.806872] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c58883-1b6c-b2c4-c945-25b8e465bb60" [ 747.806872] env[62814]: _type = "Task" [ 747.806872] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.817148] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c58883-1b6c-b2c4-c945-25b8e465bb60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.243212] env[62814]: DEBUG nova.compute.utils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 748.247205] env[62814]: DEBUG nova.compute.manager [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 748.247403] env[62814]: DEBUG nova.network.neutron [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 748.263592] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293509, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.300666] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9b04c21d-7f09-461f-aeec-3ea06d96fc8c tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.787s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 748.301105] env[62814]: DEBUG oslo_concurrency.lockutils [req-3e105ebc-24b7-4426-aa22-dfbea7ea449c req-fc389d61-7308-4931-8e84-5531c209b10d service nova] Releasing lock "refresh_cache-217e5812-95cc-4104-8d7b-82dfca2c0fcc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 748.313256] env[62814]: DEBUG nova.policy [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9df1642a53c940b2bfacc82e1ac975d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '446e63a428c34d38ad07d01f0c90b23c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 748.322971] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c58883-1b6c-b2c4-c945-25b8e465bb60, 'name': SearchDatastore_Task, 'duration_secs': 0.010719} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.323749] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 748.323749] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 217e5812-95cc-4104-8d7b-82dfca2c0fcc/217e5812-95cc-4104-8d7b-82dfca2c0fcc.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 748.323749] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-75c9467a-756b-4dfa-a2d6-cb3459b9caba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.330979] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 748.330979] env[62814]: value = "task-4293510" [ 748.330979] env[62814]: _type = "Task" [ 748.330979] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.339884] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293510, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.511530] env[62814]: DEBUG nova.network.neutron [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Successfully updated port: ce7e8021-3ebf-4e8c-a0c2-618fa44c6957 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 748.645056] env[62814]: DEBUG nova.network.neutron [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Successfully created port: 08bab1d2-6296-46f7-baf6-4344d1bbb0ef {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 748.746783] env[62814]: DEBUG nova.compute.manager [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 748.766087] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293509, 'name': ReconfigVM_Task, 'duration_secs': 1.075302} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.766840] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Reconfigured VM instance instance-00000020 to attach disk [datastore2] 369cd937-4c18-4068-ae59-70a1d585094b/369cd937-4c18-4068-ae59-70a1d585094b.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 748.767623] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abf3769d-7d29-4257-9ce5-caf8ba1e06cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.777026] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for the task: (returnval){ [ 748.777026] env[62814]: value = "task-4293511" [ 748.777026] env[62814]: _type = "Task" [ 748.777026] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.785595] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293511, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.805504] env[62814]: DEBUG nova.compute.manager [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 748.841366] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293510, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467781} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.844085] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 217e5812-95cc-4104-8d7b-82dfca2c0fcc/217e5812-95cc-4104-8d7b-82dfca2c0fcc.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 748.844314] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 748.844742] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2825ab0a-2255-45d0-bf25-4506c496f9b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.850823] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 748.850823] env[62814]: value = "task-4293512" [ 748.850823] env[62814]: _type = "Task" [ 748.850823] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.859067] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293512, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.193526] env[62814]: DEBUG nova.compute.manager [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received event network-vif-plugged-ce7e8021-3ebf-4e8c-a0c2-618fa44c6957 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 749.193698] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] Acquiring lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 749.193917] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 749.194131] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 749.194263] env[62814]: DEBUG nova.compute.manager [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] No waiting events found dispatching network-vif-plugged-ce7e8021-3ebf-4e8c-a0c2-618fa44c6957 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 749.194421] env[62814]: WARNING nova.compute.manager [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received unexpected event network-vif-plugged-ce7e8021-3ebf-4e8c-a0c2-618fa44c6957 for instance with vm_state building and task_state spawning. [ 749.194570] env[62814]: DEBUG nova.compute.manager [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received event network-changed-ce7e8021-3ebf-4e8c-a0c2-618fa44c6957 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 749.194714] env[62814]: DEBUG nova.compute.manager [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Refreshing instance network info cache due to event network-changed-ce7e8021-3ebf-4e8c-a0c2-618fa44c6957. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 749.194884] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] Acquiring lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.195281] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] Acquired lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.195481] env[62814]: DEBUG nova.network.neutron [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Refreshing network info cache for port ce7e8021-3ebf-4e8c-a0c2-618fa44c6957 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 749.290649] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293511, 'name': Rename_Task, 'duration_secs': 0.143393} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.290936] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 749.291196] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d25b9e3b-ae3e-451a-92d1-30c8943e09ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.300300] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for the task: (returnval){ [ 749.300300] env[62814]: value = "task-4293513" [ 749.300300] env[62814]: _type = "Task" [ 749.300300] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.308306] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293513, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.333081] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 749.344674] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-041696b1-8816-4f28-8c65-329e6b983a46 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.351800] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee243015-b959-4fdc-b085-c74e28e6b0c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.369674] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293512, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059102} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.397526] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 749.398590] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22404a2d-072c-45be-91a5-2e5e4ae345ac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.401661] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1b60c62-bb74-4156-8563-74b802059e3f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.417901] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca957ed0-0015-44d5-9b9f-29771af6b0d6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.430248] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] 217e5812-95cc-4104-8d7b-82dfca2c0fcc/217e5812-95cc-4104-8d7b-82dfca2c0fcc.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 749.430585] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-adb9c77b-d28d-46eb-a8eb-8012d577bf27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.455823] env[62814]: DEBUG nova.compute.provider_tree [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.458670] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 749.458670] env[62814]: value = "task-4293514" [ 749.458670] env[62814]: _type = "Task" [ 749.458670] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.466476] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293514, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.750885] env[62814]: DEBUG nova.network.neutron [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.766549] env[62814]: DEBUG nova.compute.manager [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 749.796973] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 749.796973] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.796973] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 749.797481] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.797481] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 749.797481] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 749.797481] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 749.797481] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 749.797686] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 749.797686] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 749.797686] env[62814]: DEBUG nova.virt.hardware [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 749.797996] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9c54a6-6b1e-4cc7-b2c1-64ce0508ce2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.815951] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293513, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.817779] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-005a9d4b-c6ad-4f7e-8c07-9fccd39c0284 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.881640] env[62814]: DEBUG nova.network.neutron [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.959737] env[62814]: DEBUG nova.scheduler.client.report [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 749.973108] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293514, 'name': ReconfigVM_Task, 'duration_secs': 0.283536} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.974030] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Reconfigured VM instance instance-00000021 to attach disk [datastore2] 217e5812-95cc-4104-8d7b-82dfca2c0fcc/217e5812-95cc-4104-8d7b-82dfca2c0fcc.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 749.974586] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c4ad9e3-c1f6-4596-bc45-4384d4c4178f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.980823] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 749.980823] env[62814]: value = "task-4293515" [ 749.980823] env[62814]: _type = "Task" [ 749.980823] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.989040] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293515, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.138907] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquiring lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 750.138907] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.138907] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquiring lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 750.138907] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.139293] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.140100] env[62814]: INFO nova.compute.manager [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Terminating instance [ 750.261120] env[62814]: DEBUG nova.compute.manager [req-a82678eb-e19e-4f6e-a2fa-32a93e4d930c req-c7558394-c67a-4333-bee1-25ecd6b9cdc3 service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Received event network-vif-plugged-08bab1d2-6296-46f7-baf6-4344d1bbb0ef {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 750.261250] env[62814]: DEBUG oslo_concurrency.lockutils [req-a82678eb-e19e-4f6e-a2fa-32a93e4d930c req-c7558394-c67a-4333-bee1-25ecd6b9cdc3 service nova] Acquiring lock "9343301c-dfe9-41b0-b4a0-067af544d297-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 750.261465] env[62814]: DEBUG oslo_concurrency.lockutils [req-a82678eb-e19e-4f6e-a2fa-32a93e4d930c req-c7558394-c67a-4333-bee1-25ecd6b9cdc3 service nova] Lock "9343301c-dfe9-41b0-b4a0-067af544d297-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.261672] env[62814]: DEBUG oslo_concurrency.lockutils [req-a82678eb-e19e-4f6e-a2fa-32a93e4d930c req-c7558394-c67a-4333-bee1-25ecd6b9cdc3 service nova] Lock "9343301c-dfe9-41b0-b4a0-067af544d297-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.261863] env[62814]: DEBUG nova.compute.manager [req-a82678eb-e19e-4f6e-a2fa-32a93e4d930c req-c7558394-c67a-4333-bee1-25ecd6b9cdc3 service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] No waiting events found dispatching network-vif-plugged-08bab1d2-6296-46f7-baf6-4344d1bbb0ef {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 750.262465] env[62814]: WARNING nova.compute.manager [req-a82678eb-e19e-4f6e-a2fa-32a93e4d930c req-c7558394-c67a-4333-bee1-25ecd6b9cdc3 service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Received unexpected event network-vif-plugged-08bab1d2-6296-46f7-baf6-4344d1bbb0ef for instance with vm_state building and task_state spawning. [ 750.313025] env[62814]: DEBUG oslo_vmware.api [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293513, 'name': PowerOnVM_Task, 'duration_secs': 0.83749} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.313324] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 750.313496] env[62814]: INFO nova.compute.manager [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Took 8.66 seconds to spawn the instance on the hypervisor. [ 750.313669] env[62814]: DEBUG nova.compute.manager [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 750.314458] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26bc3b25-a0d4-408a-99e8-b9e3d68ed62b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.352107] env[62814]: DEBUG nova.network.neutron [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Successfully updated port: 08bab1d2-6296-46f7-baf6-4344d1bbb0ef {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 750.384694] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b67204a-4917-4536-8e4b-6f0662c1b172 req-a1ef7f9e-f836-41fc-b0f4-490fde7c53eb service nova] Releasing lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 750.472172] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.733s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.472172] env[62814]: DEBUG nova.compute.manager [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 750.473072] env[62814]: DEBUG oslo_concurrency.lockutils [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.365s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.473419] env[62814]: DEBUG nova.objects.instance [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lazy-loading 'resources' on Instance uuid 29e69c7c-08f7-4da4-9509-02a94bf971bc {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 750.493221] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293515, 'name': Rename_Task, 'duration_secs': 0.147519} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.494209] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 750.494583] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-11341502-b987-4614-b854-7e6d6b4dd83e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.502807] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 750.502807] env[62814]: value = "task-4293516" [ 750.502807] env[62814]: _type = "Task" [ 750.502807] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.516926] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293516, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.635543] env[62814]: DEBUG nova.network.neutron [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Successfully updated port: fbfe5ed2-1472-41c4-a028-38e81d37b3fe {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 750.644513] env[62814]: DEBUG nova.compute.manager [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 750.645782] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 750.652074] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2188a700-847e-4c9e-bf4d-786e18293ead {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.660611] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 750.660611] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afd11c2a-25a0-477d-afb4-35a83ff464e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.667059] env[62814]: DEBUG oslo_vmware.api [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for the task: (returnval){ [ 750.667059] env[62814]: value = "task-4293517" [ 750.667059] env[62814]: _type = "Task" [ 750.667059] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.675692] env[62814]: DEBUG oslo_vmware.api [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293517, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.834061] env[62814]: INFO nova.compute.manager [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Took 39.13 seconds to build instance. [ 750.854983] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.855112] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 750.855280] env[62814]: DEBUG nova.network.neutron [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 750.976683] env[62814]: DEBUG nova.compute.utils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 750.978629] env[62814]: DEBUG nova.compute.manager [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 750.978811] env[62814]: DEBUG nova.network.neutron [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 751.015861] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293516, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.066761] env[62814]: DEBUG nova.policy [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '275a668099f546aa9e4f1e9cc165cbfe', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1a4f27dd44c34eb5814db7e462dd2373', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 751.179427] env[62814]: DEBUG oslo_vmware.api [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293517, 'name': PowerOffVM_Task, 'duration_secs': 0.328114} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.180944] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 751.180944] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 751.182050] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3dedd791-1a6b-43ff-bedf-f85e3e2414b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.249018] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 751.249361] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 751.249670] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Deleting the datastore file [datastore2] a589a3d8-20dc-4ff5-a192-c540e29f39d6 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 751.250394] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b277388-14f4-40ad-b5f7-56e0b4424764 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.263040] env[62814]: DEBUG oslo_vmware.api [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for the task: (returnval){ [ 751.263040] env[62814]: value = "task-4293519" [ 751.263040] env[62814]: _type = "Task" [ 751.263040] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.266060] env[62814]: DEBUG nova.compute.manager [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received event network-vif-plugged-fbfe5ed2-1472-41c4-a028-38e81d37b3fe {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 751.266353] env[62814]: DEBUG oslo_concurrency.lockutils [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] Acquiring lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.266642] env[62814]: DEBUG oslo_concurrency.lockutils [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.266951] env[62814]: DEBUG oslo_concurrency.lockutils [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.267117] env[62814]: DEBUG nova.compute.manager [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] No waiting events found dispatching network-vif-plugged-fbfe5ed2-1472-41c4-a028-38e81d37b3fe {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 751.267479] env[62814]: WARNING nova.compute.manager [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received unexpected event network-vif-plugged-fbfe5ed2-1472-41c4-a028-38e81d37b3fe for instance with vm_state building and task_state spawning. [ 751.267884] env[62814]: DEBUG nova.compute.manager [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received event network-changed-fbfe5ed2-1472-41c4-a028-38e81d37b3fe {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 751.267884] env[62814]: DEBUG nova.compute.manager [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Refreshing instance network info cache due to event network-changed-fbfe5ed2-1472-41c4-a028-38e81d37b3fe. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 751.268198] env[62814]: DEBUG oslo_concurrency.lockutils [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] Acquiring lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.268439] env[62814]: DEBUG oslo_concurrency.lockutils [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] Acquired lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 751.268675] env[62814]: DEBUG nova.network.neutron [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Refreshing network info cache for port fbfe5ed2-1472-41c4-a028-38e81d37b3fe {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 751.285918] env[62814]: DEBUG oslo_vmware.api [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293519, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.337114] env[62814]: DEBUG oslo_concurrency.lockutils [None req-35e0408d-2220-4c1c-a706-b4a7667e1fe9 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Lock "369cd937-4c18-4068-ae59-70a1d585094b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.817s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.435018] env[62814]: DEBUG nova.network.neutron [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.486112] env[62814]: DEBUG nova.network.neutron [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Successfully created port: 2c623055-7773-4e8b-86c0-8765158b3620 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 751.486112] env[62814]: DEBUG nova.compute.manager [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 751.510454] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17855a69-9fe5-4096-bae8-92d0f7acbc5a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.518981] env[62814]: DEBUG oslo_vmware.api [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293516, 'name': PowerOnVM_Task, 'duration_secs': 0.516262} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.521821] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 751.522325] env[62814]: INFO nova.compute.manager [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Took 7.23 seconds to spawn the instance on the hypervisor. [ 751.522657] env[62814]: DEBUG nova.compute.manager [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 751.523758] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7550b38e-9cac-4a2d-9ece-c083b0fbddbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.527653] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1b0ad2-ba7a-4ce8-a257-e9b83a43879c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.575124] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfb98e6-91fb-454e-a82a-5857804d5927 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.585936] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920f0001-bd6b-4024-952f-14ad464e42b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.600505] env[62814]: DEBUG nova.compute.provider_tree [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.669163] env[62814]: DEBUG nova.network.neutron [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance_info_cache with network_info: [{"id": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "address": "fa:16:3e:db:ee:dc", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08bab1d2-62", "ovs_interfaceid": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.780414] env[62814]: DEBUG oslo_vmware.api [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Task: {'id': task-4293519, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.50323} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.780856] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 751.781123] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 751.781364] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 751.782857] env[62814]: INFO nova.compute.manager [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Took 1.14 seconds to destroy the instance on the hypervisor. [ 751.783213] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 751.784253] env[62814]: DEBUG nova.compute.manager [-] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 751.784363] env[62814]: DEBUG nova.network.neutron [-] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 751.839070] env[62814]: DEBUG nova.compute.manager [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 751.854137] env[62814]: DEBUG nova.network.neutron [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.051524] env[62814]: DEBUG nova.network.neutron [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.081362] env[62814]: INFO nova.compute.manager [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Took 39.96 seconds to build instance. [ 752.103963] env[62814]: DEBUG nova.scheduler.client.report [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 752.174265] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 752.174640] env[62814]: DEBUG nova.compute.manager [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Instance network_info: |[{"id": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "address": "fa:16:3e:db:ee:dc", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08bab1d2-62", "ovs_interfaceid": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 752.175086] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:db:ee:dc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b83383f-ed7a-4efd-aef7-aa8c15649d07', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '08bab1d2-6296-46f7-baf6-4344d1bbb0ef', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 752.182756] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating folder: Project (446e63a428c34d38ad07d01f0c90b23c). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 752.183060] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-caa62c73-329b-40f3-ae24-0aaebf9f9139 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.195568] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Created folder: Project (446e63a428c34d38ad07d01f0c90b23c) in parent group-v845547. [ 752.195568] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating folder: Instances. Parent ref: group-v845641. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 752.195568] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c9db4393-459f-4269-9277-f94a08f5af60 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.207788] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Created folder: Instances in parent group-v845641. [ 752.208042] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 752.208244] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 752.208797] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa26a189-f5e3-4217-aa5a-e02fb050c7ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.232650] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 752.232650] env[62814]: value = "task-4293522" [ 752.232650] env[62814]: _type = "Task" [ 752.232650] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.243828] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293522, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.364206] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 752.501810] env[62814]: DEBUG nova.compute.manager [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 752.531445] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 752.531644] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.531770] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 752.531930] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.532094] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 752.532246] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 752.532458] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 752.532616] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 752.532778] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 752.532974] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 752.533149] env[62814]: DEBUG nova.virt.hardware [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 752.534033] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8efef8d-fa30-41ef-b5c0-8d815f839f60 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.542815] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fecbc720-09f4-408b-b8ff-a4a73a3552c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.559557] env[62814]: DEBUG oslo_concurrency.lockutils [req-120cb0c2-7319-4024-9622-ac8aa01d8f92 req-05671204-9e76-4492-8522-0aa422a744dd service nova] Releasing lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 752.583252] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b1a09ce7-5a76-4084-a80a-deb3ef15deda tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.324s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 752.611572] env[62814]: DEBUG oslo_concurrency.lockutils [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.138s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 752.615026] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.479s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.616354] env[62814]: INFO nova.compute.claims [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.638866] env[62814]: INFO nova.scheduler.client.report [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Deleted allocations for instance 29e69c7c-08f7-4da4-9509-02a94bf971bc [ 752.666879] env[62814]: DEBUG nova.compute.manager [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Received event network-changed-08bab1d2-6296-46f7-baf6-4344d1bbb0ef {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 752.669486] env[62814]: DEBUG nova.compute.manager [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Refreshing instance network info cache due to event network-changed-08bab1d2-6296-46f7-baf6-4344d1bbb0ef. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 752.669486] env[62814]: DEBUG oslo_concurrency.lockutils [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] Acquiring lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.669486] env[62814]: DEBUG oslo_concurrency.lockutils [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] Acquired lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 752.669780] env[62814]: DEBUG nova.network.neutron [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Refreshing network info cache for port 08bab1d2-6296-46f7-baf6-4344d1bbb0ef {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 752.726897] env[62814]: DEBUG nova.network.neutron [-] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.747744] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293522, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.879862] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f30df30-3ef3-4c19-828c-f5a0dd01f2a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.887468] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-835936e1-c8b9-4699-9128-e98edef85992 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Suspending the VM {{(pid=62814) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 752.887757] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-dac1a551-7862-46d2-a62f-fa9747720d96 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.896682] env[62814]: DEBUG oslo_vmware.api [None req-835936e1-c8b9-4699-9128-e98edef85992 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 752.896682] env[62814]: value = "task-4293523" [ 752.896682] env[62814]: _type = "Task" [ 752.896682] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.907247] env[62814]: DEBUG oslo_vmware.api [None req-835936e1-c8b9-4699-9128-e98edef85992 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293523, 'name': SuspendVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.004789] env[62814]: DEBUG nova.network.neutron [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Successfully updated port: 92958ac5-6305-4ce3-aa70-0c6080099c48 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 753.088609] env[62814]: DEBUG nova.compute.manager [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 753.149264] env[62814]: DEBUG oslo_concurrency.lockutils [None req-776cd291-7448-4539-a7a1-c49d74027a4e tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "29e69c7c-08f7-4da4-9509-02a94bf971bc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.526s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.232988] env[62814]: INFO nova.compute.manager [-] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Took 1.45 seconds to deallocate network for instance. [ 753.254176] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293522, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.272100] env[62814]: DEBUG nova.network.neutron [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Successfully updated port: 2c623055-7773-4e8b-86c0-8765158b3620 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 753.408613] env[62814]: DEBUG oslo_vmware.api [None req-835936e1-c8b9-4699-9128-e98edef85992 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293523, 'name': SuspendVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.512395] env[62814]: DEBUG nova.network.neutron [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updated VIF entry in instance network info cache for port 08bab1d2-6296-46f7-baf6-4344d1bbb0ef. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 753.512856] env[62814]: DEBUG nova.network.neutron [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance_info_cache with network_info: [{"id": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "address": "fa:16:3e:db:ee:dc", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08bab1d2-62", "ovs_interfaceid": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.514433] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.514543] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquired lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.514685] env[62814]: DEBUG nova.network.neutron [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.615838] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.750944] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.752029] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293522, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.767264] env[62814]: DEBUG nova.compute.manager [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Received event network-changed-efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 753.767264] env[62814]: DEBUG nova.compute.manager [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Refreshing instance network info cache due to event network-changed-efe64bba-a178-4079-b256-51b9e74293c3. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 753.767264] env[62814]: DEBUG oslo_concurrency.lockutils [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] Acquiring lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.767264] env[62814]: DEBUG oslo_concurrency.lockutils [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] Acquired lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.768196] env[62814]: DEBUG nova.network.neutron [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Refreshing network info cache for port efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 753.776970] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "refresh_cache-56f9e03c-1157-442f-8add-156627a6ec1e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.777153] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquired lock "refresh_cache-56f9e03c-1157-442f-8add-156627a6ec1e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.777383] env[62814]: DEBUG nova.network.neutron [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.909812] env[62814]: DEBUG oslo_vmware.api [None req-835936e1-c8b9-4699-9128-e98edef85992 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293523, 'name': SuspendVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.020351] env[62814]: DEBUG oslo_concurrency.lockutils [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] Releasing lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 754.020680] env[62814]: DEBUG nova.compute.manager [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Received event network-vif-deleted-888abe76-da04-4e32-9f81-159cde73dbee {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 754.020874] env[62814]: INFO nova.compute.manager [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Neutron deleted interface 888abe76-da04-4e32-9f81-159cde73dbee; detaching it from the instance and deleting it from the info cache [ 754.021060] env[62814]: DEBUG nova.network.neutron [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.074930] env[62814]: DEBUG nova.network.neutron [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.193192] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6780336a-d3a1-4fb8-a7cd-e4e482ca3870 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.200815] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2f5e81-ad39-4aff-99fc-c5cf48d5c9a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.246605] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e0a591-a5ef-42dc-8903-575314f11c4c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.254284] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293522, 'name': CreateVM_Task, 'duration_secs': 1.57851} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.258297] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 754.259203] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.259689] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.260105] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 754.261981] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085949e1-de37-466e-8a39-8e1260d891b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.268780] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bcd195bf-aa47-4093-9e64-6704d6ac1c0a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.286982] env[62814]: DEBUG nova.compute.provider_tree [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.291837] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 754.291837] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5261a563-3f4b-f250-2966-cee9c375ea37" [ 754.291837] env[62814]: _type = "Task" [ 754.291837] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.303401] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5261a563-3f4b-f250-2966-cee9c375ea37, 'name': SearchDatastore_Task, 'duration_secs': 0.010263} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.303755] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 754.304099] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.304428] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.304621] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.304815] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 754.305229] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40e1a030-c4e8-44f6-884b-eb1a85e7aa06 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.315037] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 754.315240] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 754.316262] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5da1417-c655-438b-bf57-f1acbb29f20d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.321489] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 754.321489] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f4536-df9f-a891-110b-e36f8c00cd68" [ 754.321489] env[62814]: _type = "Task" [ 754.321489] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.331609] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f4536-df9f-a891-110b-e36f8c00cd68, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.347341] env[62814]: DEBUG nova.network.neutron [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.410513] env[62814]: DEBUG oslo_vmware.api [None req-835936e1-c8b9-4699-9128-e98edef85992 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293523, 'name': SuspendVM_Task} progress is 62%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.526805] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd885a7e-e1da-4647-8591-5801734cdf3f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.536799] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a797862f-5184-4f46-9d9f-9ae673d36869 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.579136] env[62814]: DEBUG nova.compute.manager [req-b941a2c1-4a60-4c66-8948-a9b50a8539e3 req-3a14b00d-98bd-4dff-a637-67943ceba46a service nova] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Detach interface failed, port_id=888abe76-da04-4e32-9f81-159cde73dbee, reason: Instance a589a3d8-20dc-4ff5-a192-c540e29f39d6 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 754.702857] env[62814]: DEBUG nova.network.neutron [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Updating instance_info_cache with network_info: [{"id": "2c623055-7773-4e8b-86c0-8765158b3620", "address": "fa:16:3e:f9:6b:6b", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.109", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c623055-77", "ovs_interfaceid": "2c623055-7773-4e8b-86c0-8765158b3620", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.800513] env[62814]: DEBUG nova.scheduler.client.report [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 754.834968] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f4536-df9f-a891-110b-e36f8c00cd68, 'name': SearchDatastore_Task, 'duration_secs': 0.008577} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.835885] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec81b67a-f09e-4de2-9650-f638760dc7a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.842565] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 754.842565] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a6d096-1a23-6a30-4f15-8a15fe074aba" [ 754.842565] env[62814]: _type = "Task" [ 754.842565] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.853679] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a6d096-1a23-6a30-4f15-8a15fe074aba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.875161] env[62814]: DEBUG nova.network.neutron [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Updated VIF entry in instance network info cache for port efe64bba-a178-4079-b256-51b9e74293c3. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 754.875543] env[62814]: DEBUG nova.network.neutron [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Updating instance_info_cache with network_info: [{"id": "efe64bba-a178-4079-b256-51b9e74293c3", "address": "fa:16:3e:fe:f6:d8", "network": {"id": "755078a7-2f0f-43e8-b6a3-ca1a183cad9a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-249930933-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85eefae1aad42d8bed6bed203b8d221", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a78d5760-0bb1-4476-9578-8ad3c3144439", "external-id": "nsx-vlan-transportzone-325", "segmentation_id": 325, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapefe64bba-a1", "ovs_interfaceid": "efe64bba-a178-4079-b256-51b9e74293c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.915055] env[62814]: DEBUG oslo_vmware.api [None req-835936e1-c8b9-4699-9128-e98edef85992 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293523, 'name': SuspendVM_Task, 'duration_secs': 1.64217} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.915055] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-835936e1-c8b9-4699-9128-e98edef85992 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Suspended the VM {{(pid=62814) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 754.915055] env[62814]: DEBUG nova.compute.manager [None req-835936e1-c8b9-4699-9128-e98edef85992 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 754.915344] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-133eac00-e21a-48a7-9d83-fb862ee6b7ad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.981468] env[62814]: DEBUG nova.network.neutron [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Updating instance_info_cache with network_info: [{"id": "ce7e8021-3ebf-4e8c-a0c2-618fa44c6957", "address": "fa:16:3e:2d:0b:38", "network": {"id": "4a0aa60f-1f50-4cc4-b64e-96e75cfc8d42", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1969588786", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce7e8021-3e", "ovs_interfaceid": "ce7e8021-3ebf-4e8c-a0c2-618fa44c6957", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fbfe5ed2-1472-41c4-a028-38e81d37b3fe", "address": "fa:16:3e:1c:21:12", "network": {"id": "4fa67c69-1661-4203-a301-da311d1b4489", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-824717230", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.25", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbfe5ed2-14", "ovs_interfaceid": "fbfe5ed2-1472-41c4-a028-38e81d37b3fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "92958ac5-6305-4ce3-aa70-0c6080099c48", "address": "fa:16:3e:26:80:7a", "network": {"id": "4a0aa60f-1f50-4cc4-b64e-96e75cfc8d42", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1969588786", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.106", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92958ac5-63", "ovs_interfaceid": "92958ac5-6305-4ce3-aa70-0c6080099c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.056549] env[62814]: DEBUG nova.compute.manager [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received event network-vif-plugged-92958ac5-6305-4ce3-aa70-0c6080099c48 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 755.056774] env[62814]: DEBUG oslo_concurrency.lockutils [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] Acquiring lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.056984] env[62814]: DEBUG oslo_concurrency.lockutils [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.057163] env[62814]: DEBUG oslo_concurrency.lockutils [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.057358] env[62814]: DEBUG nova.compute.manager [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] No waiting events found dispatching network-vif-plugged-92958ac5-6305-4ce3-aa70-0c6080099c48 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 755.057524] env[62814]: WARNING nova.compute.manager [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received unexpected event network-vif-plugged-92958ac5-6305-4ce3-aa70-0c6080099c48 for instance with vm_state building and task_state spawning. [ 755.057731] env[62814]: DEBUG nova.compute.manager [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received event network-changed-92958ac5-6305-4ce3-aa70-0c6080099c48 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 755.057817] env[62814]: DEBUG nova.compute.manager [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Refreshing instance network info cache due to event network-changed-92958ac5-6305-4ce3-aa70-0c6080099c48. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 755.058286] env[62814]: DEBUG oslo_concurrency.lockutils [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] Acquiring lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.205178] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Releasing lock "refresh_cache-56f9e03c-1157-442f-8add-156627a6ec1e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 755.205659] env[62814]: DEBUG nova.compute.manager [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Instance network_info: |[{"id": "2c623055-7773-4e8b-86c0-8765158b3620", "address": "fa:16:3e:f9:6b:6b", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.109", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c623055-77", "ovs_interfaceid": "2c623055-7773-4e8b-86c0-8765158b3620", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 755.206214] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:6b:6b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2c623055-7773-4e8b-86c0-8765158b3620', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 755.216522] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 755.216811] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 755.217109] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d2a9c6a-52ce-4bc1-8195-1191371d4cc3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.242542] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 755.242542] env[62814]: value = "task-4293524" [ 755.242542] env[62814]: _type = "Task" [ 755.242542] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.250362] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293524, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.310041] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.693s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.310041] env[62814]: DEBUG nova.compute.manager [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 755.313892] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.664s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.318042] env[62814]: INFO nova.compute.claims [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 755.355324] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a6d096-1a23-6a30-4f15-8a15fe074aba, 'name': SearchDatastore_Task, 'duration_secs': 0.009547} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.355324] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 755.355324] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 9343301c-dfe9-41b0-b4a0-067af544d297/9343301c-dfe9-41b0-b4a0-067af544d297.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 755.355324] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-07e691de-87f0-494d-affa-5af8847d84bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.362947] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 755.362947] env[62814]: value = "task-4293525" [ 755.362947] env[62814]: _type = "Task" [ 755.362947] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.372320] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293525, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.381393] env[62814]: DEBUG oslo_concurrency.lockutils [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] Releasing lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 755.381592] env[62814]: DEBUG nova.compute.manager [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Received event network-vif-plugged-2c623055-7773-4e8b-86c0-8765158b3620 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 755.381809] env[62814]: DEBUG oslo_concurrency.lockutils [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] Acquiring lock "56f9e03c-1157-442f-8add-156627a6ec1e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.382054] env[62814]: DEBUG oslo_concurrency.lockutils [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] Lock "56f9e03c-1157-442f-8add-156627a6ec1e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.382239] env[62814]: DEBUG oslo_concurrency.lockutils [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] Lock "56f9e03c-1157-442f-8add-156627a6ec1e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.382396] env[62814]: DEBUG nova.compute.manager [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] No waiting events found dispatching network-vif-plugged-2c623055-7773-4e8b-86c0-8765158b3620 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 755.382611] env[62814]: WARNING nova.compute.manager [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Received unexpected event network-vif-plugged-2c623055-7773-4e8b-86c0-8765158b3620 for instance with vm_state building and task_state spawning. [ 755.382774] env[62814]: DEBUG nova.compute.manager [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Received event network-changed-2c623055-7773-4e8b-86c0-8765158b3620 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 755.382937] env[62814]: DEBUG nova.compute.manager [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Refreshing instance network info cache due to event network-changed-2c623055-7773-4e8b-86c0-8765158b3620. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 755.383130] env[62814]: DEBUG oslo_concurrency.lockutils [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] Acquiring lock "refresh_cache-56f9e03c-1157-442f-8add-156627a6ec1e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.383276] env[62814]: DEBUG oslo_concurrency.lockutils [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] Acquired lock "refresh_cache-56f9e03c-1157-442f-8add-156627a6ec1e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.383444] env[62814]: DEBUG nova.network.neutron [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Refreshing network info cache for port 2c623055-7773-4e8b-86c0-8765158b3620 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 755.484085] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Releasing lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 755.484579] env[62814]: DEBUG nova.compute.manager [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Instance network_info: |[{"id": "ce7e8021-3ebf-4e8c-a0c2-618fa44c6957", "address": "fa:16:3e:2d:0b:38", "network": {"id": "4a0aa60f-1f50-4cc4-b64e-96e75cfc8d42", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1969588786", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce7e8021-3e", "ovs_interfaceid": "ce7e8021-3ebf-4e8c-a0c2-618fa44c6957", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fbfe5ed2-1472-41c4-a028-38e81d37b3fe", "address": "fa:16:3e:1c:21:12", "network": {"id": "4fa67c69-1661-4203-a301-da311d1b4489", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-824717230", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.25", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbfe5ed2-14", "ovs_interfaceid": "fbfe5ed2-1472-41c4-a028-38e81d37b3fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "92958ac5-6305-4ce3-aa70-0c6080099c48", "address": "fa:16:3e:26:80:7a", "network": {"id": "4a0aa60f-1f50-4cc4-b64e-96e75cfc8d42", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1969588786", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.106", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92958ac5-63", "ovs_interfaceid": "92958ac5-6305-4ce3-aa70-0c6080099c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 755.484942] env[62814]: DEBUG oslo_concurrency.lockutils [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] Acquired lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.485190] env[62814]: DEBUG nova.network.neutron [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Refreshing network info cache for port 92958ac5-6305-4ce3-aa70-0c6080099c48 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 755.489600] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2d:0b:38', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ce7e8021-3ebf-4e8c-a0c2-618fa44c6957', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:21:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50886eea-591a-452c-a27b-5f22cfc9df85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fbfe5ed2-1472-41c4-a028-38e81d37b3fe', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:80:7a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92958ac5-6305-4ce3-aa70-0c6080099c48', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 755.501539] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Creating folder: Project (d14adacd89bc460d8d1a7c7d2ecc5d7f). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 755.502957] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5e8974e2-db13-4251-8db9-0be9d6dbee5c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.523027] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Created folder: Project (d14adacd89bc460d8d1a7c7d2ecc5d7f) in parent group-v845547. [ 755.523027] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Creating folder: Instances. Parent ref: group-v845645. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 755.523027] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ddb00217-692c-4a7e-8bbd-e24a94cc9479 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.533920] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Created folder: Instances in parent group-v845645. [ 755.534229] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 755.534445] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 755.534660] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e48983c-45de-41b3-b130-92ad960d1ae3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.566782] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 755.566782] env[62814]: value = "task-4293528" [ 755.566782] env[62814]: _type = "Task" [ 755.566782] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.580028] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293528, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.757079] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293524, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.826599] env[62814]: DEBUG nova.compute.utils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 755.828644] env[62814]: DEBUG nova.compute.manager [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 755.828777] env[62814]: DEBUG nova.network.neutron [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 755.879353] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293525, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.905548] env[62814]: DEBUG nova.policy [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd0503b34d08a4c0d9eb371231b0135f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0e8c68a04eda4221925a5397578c0280', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 756.080351] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293528, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.236407] env[62814]: DEBUG nova.network.neutron [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Updated VIF entry in instance network info cache for port 2c623055-7773-4e8b-86c0-8765158b3620. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 756.236407] env[62814]: DEBUG nova.network.neutron [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Updating instance_info_cache with network_info: [{"id": "2c623055-7773-4e8b-86c0-8765158b3620", "address": "fa:16:3e:f9:6b:6b", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.109", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2c623055-77", "ovs_interfaceid": "2c623055-7773-4e8b-86c0-8765158b3620", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.264340] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293524, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.318411] env[62814]: DEBUG nova.network.neutron [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Updated VIF entry in instance network info cache for port 92958ac5-6305-4ce3-aa70-0c6080099c48. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 756.318411] env[62814]: DEBUG nova.network.neutron [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Updating instance_info_cache with network_info: [{"id": "ce7e8021-3ebf-4e8c-a0c2-618fa44c6957", "address": "fa:16:3e:2d:0b:38", "network": {"id": "4a0aa60f-1f50-4cc4-b64e-96e75cfc8d42", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1969588786", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce7e8021-3e", "ovs_interfaceid": "ce7e8021-3ebf-4e8c-a0c2-618fa44c6957", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fbfe5ed2-1472-41c4-a028-38e81d37b3fe", "address": "fa:16:3e:1c:21:12", "network": {"id": "4fa67c69-1661-4203-a301-da311d1b4489", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-824717230", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.25", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbfe5ed2-14", "ovs_interfaceid": "fbfe5ed2-1472-41c4-a028-38e81d37b3fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "92958ac5-6305-4ce3-aa70-0c6080099c48", "address": "fa:16:3e:26:80:7a", "network": {"id": "4a0aa60f-1f50-4cc4-b64e-96e75cfc8d42", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1969588786", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.106", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92958ac5-63", "ovs_interfaceid": "92958ac5-6305-4ce3-aa70-0c6080099c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.335338] env[62814]: DEBUG nova.compute.manager [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 756.376270] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293525, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523402} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.378420] env[62814]: DEBUG nova.network.neutron [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Successfully created port: 02f411b5-731e-45d9-a7dc-9aca642cd9aa {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.379976] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 9343301c-dfe9-41b0-b4a0-067af544d297/9343301c-dfe9-41b0-b4a0-067af544d297.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 756.382048] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 756.382048] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ab9e97f-47f7-4ae6-8450-8375ddd5d7af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.389305] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 756.389305] env[62814]: value = "task-4293529" [ 756.389305] env[62814]: _type = "Task" [ 756.389305] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.398592] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.587935] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293528, 'name': CreateVM_Task, 'duration_secs': 0.82937} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.588555] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 756.589493] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.590521] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 756.590990] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 756.591522] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fb0a42e-1a7f-458f-bb4e-86a8876472ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.597371] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 756.597371] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5297ebe0-ea85-bb59-20a4-031150767ec6" [ 756.597371] env[62814]: _type = "Task" [ 756.597371] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.608876] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5297ebe0-ea85-bb59-20a4-031150767ec6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.741494] env[62814]: DEBUG oslo_concurrency.lockutils [req-0cde3cdd-2b48-4409-a1b1-89beeb9a934b req-ad5f2df9-07d4-4657-a7ad-ffbf9456c868 service nova] Releasing lock "refresh_cache-56f9e03c-1157-442f-8add-156627a6ec1e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 756.755258] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293524, 'name': CreateVM_Task, 'duration_secs': 1.454977} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.757856] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 756.758848] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.823552] env[62814]: DEBUG oslo_concurrency.lockutils [req-de685813-12a3-4bea-a36f-bb99d341ef01 req-279d0518-3336-4e78-9367-4cb2373e338a service nova] Releasing lock "refresh_cache-961dedc6-5ddd-4620-bfa7-54cf0c520af3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 756.902839] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070486} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.902839] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 756.903788] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a990165-82fb-4d70-ae10-b7e79dba96ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.926972] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] 9343301c-dfe9-41b0-b4a0-067af544d297/9343301c-dfe9-41b0-b4a0-067af544d297.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.930344] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80855a5a-1369-4f6d-a9f9-8b94ef6d8c62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.950713] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 756.950713] env[62814]: value = "task-4293530" [ 756.950713] env[62814]: _type = "Task" [ 756.950713] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.955337] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0804f162-59ea-4367-a1f2-05925f4bb053 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.962923] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293530, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.966540] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b596da-6e8a-489c-add2-eee5495a6ca9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.998805] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436b027e-1c95-4c44-a322-b30e0ff42b5b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.006486] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b45798-a0bc-45e0-a50b-2e5757fc5ccf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.023619] env[62814]: DEBUG nova.compute.provider_tree [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 757.107924] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5297ebe0-ea85-bb59-20a4-031150767ec6, 'name': SearchDatastore_Task, 'duration_secs': 0.011457} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.108276] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.108528] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 757.108795] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.108988] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.109213] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 757.109536] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.109858] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 757.110120] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-db97dd75-c328-487f-8286-f838ba3aa3de {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.112103] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-deeb6b77-5623-4390-b4ea-2fcfe6627f92 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.121441] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 757.121441] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527f3301-ac81-fd8f-f4ff-c88a5a8f9545" [ 757.121441] env[62814]: _type = "Task" [ 757.121441] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.128301] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 757.128431] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 757.129475] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c9c6299-42e9-4464-ad30-1452dd86a018 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.135464] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527f3301-ac81-fd8f-f4ff-c88a5a8f9545, 'name': SearchDatastore_Task, 'duration_secs': 0.009121} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.138017] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 757.138017] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 757.138017] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.138658] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 757.138658] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5204db4b-91fd-52c1-9f74-10c01674c199" [ 757.138658] env[62814]: _type = "Task" [ 757.138658] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.147689] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5204db4b-91fd-52c1-9f74-10c01674c199, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.229070] env[62814]: DEBUG nova.compute.manager [req-33a1bc8d-2cba-4fbd-88b9-440600cd9f98 req-1819cda6-4a00-4ab5-a86d-8b607e2192b6 service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Received event network-changed-efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 757.229249] env[62814]: DEBUG nova.compute.manager [req-33a1bc8d-2cba-4fbd-88b9-440600cd9f98 req-1819cda6-4a00-4ab5-a86d-8b607e2192b6 service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Refreshing instance network info cache due to event network-changed-efe64bba-a178-4079-b256-51b9e74293c3. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 757.229573] env[62814]: DEBUG oslo_concurrency.lockutils [req-33a1bc8d-2cba-4fbd-88b9-440600cd9f98 req-1819cda6-4a00-4ab5-a86d-8b607e2192b6 service nova] Acquiring lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.229735] env[62814]: DEBUG oslo_concurrency.lockutils [req-33a1bc8d-2cba-4fbd-88b9-440600cd9f98 req-1819cda6-4a00-4ab5-a86d-8b607e2192b6 service nova] Acquired lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 757.229896] env[62814]: DEBUG nova.network.neutron [req-33a1bc8d-2cba-4fbd-88b9-440600cd9f98 req-1819cda6-4a00-4ab5-a86d-8b607e2192b6 service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Refreshing network info cache for port efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 757.345425] env[62814]: DEBUG nova.compute.manager [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 757.367115] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 757.367348] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.367948] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 757.367948] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.367948] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 757.368127] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 757.368291] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 757.368454] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 757.368621] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 757.368784] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 757.368956] env[62814]: DEBUG nova.virt.hardware [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 757.370220] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b612ee5f-adab-4481-ba80-ee97dee9db31 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.378253] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb5980c8-5b3a-48c6-b0fe-d4fc0e4efb79 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.460441] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293530, 'name': ReconfigVM_Task, 'duration_secs': 0.289077} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.461221] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Reconfigured VM instance instance-00000023 to attach disk [datastore2] 9343301c-dfe9-41b0-b4a0-067af544d297/9343301c-dfe9-41b0-b4a0-067af544d297.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 757.461618] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c91ad732-4f8b-41c0-acc1-adb80539cbf4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.467804] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 757.467804] env[62814]: value = "task-4293531" [ 757.467804] env[62814]: _type = "Task" [ 757.467804] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.477962] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293531, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.544286] env[62814]: ERROR nova.scheduler.client.report [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [req-e75ee855-94e1-4952-9a1c-cae4d2c80b1f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e75ee855-94e1-4952-9a1c-cae4d2c80b1f"}]} [ 757.569761] env[62814]: DEBUG nova.scheduler.client.report [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 757.588034] env[62814]: DEBUG nova.scheduler.client.report [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 757.588357] env[62814]: DEBUG nova.compute.provider_tree [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 757.597929] env[62814]: DEBUG nova.compute.manager [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 757.597929] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.597929] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 757.598706] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be702c79-5b43-4520-ac68-64791224e517 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.609250] env[62814]: DEBUG nova.scheduler.client.report [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 757.632919] env[62814]: DEBUG nova.scheduler.client.report [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 757.655167] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5204db4b-91fd-52c1-9f74-10c01674c199, 'name': SearchDatastore_Task, 'duration_secs': 0.008489} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.655167] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55335b93-8e3d-4899-b141-bc8ab3ef96ce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.663974] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 757.663974] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7c682-8751-5a98-975b-4e13d5b8d3b9" [ 757.663974] env[62814]: _type = "Task" [ 757.663974] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.673251] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7c682-8751-5a98-975b-4e13d5b8d3b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.983248] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293531, 'name': Rename_Task, 'duration_secs': 0.165327} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.983248] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 757.983248] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6f392df8-f22c-4e7a-8842-c2ee53707482 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.989387] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 757.989387] env[62814]: value = "task-4293532" [ 757.989387] env[62814]: _type = "Task" [ 757.989387] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.005814] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293532, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.060488] env[62814]: DEBUG nova.compute.manager [req-5c8fd854-e313-4b10-966f-7f8b83d23a8c req-cddd01c4-4dd8-4a2e-a912-91efe73ac2c4 service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Received event network-vif-plugged-02f411b5-731e-45d9-a7dc-9aca642cd9aa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 758.060704] env[62814]: DEBUG oslo_concurrency.lockutils [req-5c8fd854-e313-4b10-966f-7f8b83d23a8c req-cddd01c4-4dd8-4a2e-a912-91efe73ac2c4 service nova] Acquiring lock "1a16bd56-0992-4bec-bd3f-2836dfe68579-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.060913] env[62814]: DEBUG oslo_concurrency.lockutils [req-5c8fd854-e313-4b10-966f-7f8b83d23a8c req-cddd01c4-4dd8-4a2e-a912-91efe73ac2c4 service nova] Lock "1a16bd56-0992-4bec-bd3f-2836dfe68579-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.061103] env[62814]: DEBUG oslo_concurrency.lockutils [req-5c8fd854-e313-4b10-966f-7f8b83d23a8c req-cddd01c4-4dd8-4a2e-a912-91efe73ac2c4 service nova] Lock "1a16bd56-0992-4bec-bd3f-2836dfe68579-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 758.061269] env[62814]: DEBUG nova.compute.manager [req-5c8fd854-e313-4b10-966f-7f8b83d23a8c req-cddd01c4-4dd8-4a2e-a912-91efe73ac2c4 service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] No waiting events found dispatching network-vif-plugged-02f411b5-731e-45d9-a7dc-9aca642cd9aa {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 758.061436] env[62814]: WARNING nova.compute.manager [req-5c8fd854-e313-4b10-966f-7f8b83d23a8c req-cddd01c4-4dd8-4a2e-a912-91efe73ac2c4 service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Received unexpected event network-vif-plugged-02f411b5-731e-45d9-a7dc-9aca642cd9aa for instance with vm_state building and task_state spawning. [ 758.113432] env[62814]: INFO nova.compute.manager [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] instance snapshotting [ 758.113707] env[62814]: WARNING nova.compute.manager [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 758.117482] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c79b98-f507-40a5-bfda-3bba915dcc20 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.146679] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a06296-2dd7-4427-8727-da193522d27a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.151128] env[62814]: DEBUG nova.network.neutron [req-33a1bc8d-2cba-4fbd-88b9-440600cd9f98 req-1819cda6-4a00-4ab5-a86d-8b607e2192b6 service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Updated VIF entry in instance network info cache for port efe64bba-a178-4079-b256-51b9e74293c3. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 758.151503] env[62814]: DEBUG nova.network.neutron [req-33a1bc8d-2cba-4fbd-88b9-440600cd9f98 req-1819cda6-4a00-4ab5-a86d-8b607e2192b6 service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Updating instance_info_cache with network_info: [{"id": "efe64bba-a178-4079-b256-51b9e74293c3", "address": "fa:16:3e:fe:f6:d8", "network": {"id": "755078a7-2f0f-43e8-b6a3-ca1a183cad9a", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-249930933-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f85eefae1aad42d8bed6bed203b8d221", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a78d5760-0bb1-4476-9578-8ad3c3144439", "external-id": "nsx-vlan-transportzone-325", "segmentation_id": 325, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapefe64bba-a1", "ovs_interfaceid": "efe64bba-a178-4079-b256-51b9e74293c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.174721] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7c682-8751-5a98-975b-4e13d5b8d3b9, 'name': SearchDatastore_Task, 'duration_secs': 0.017129} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.175019] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.176027] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 961dedc6-5ddd-4620-bfa7-54cf0c520af3/961dedc6-5ddd-4620-bfa7-54cf0c520af3.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 758.176027] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.176027] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 758.176027] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cffe1dca-b58c-4c8e-bc66-ceda9da99ef4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.180473] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-61536636-f120-4b93-87f6-b890e5af8a70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.189092] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 758.189092] env[62814]: value = "task-4293533" [ 758.189092] env[62814]: _type = "Task" [ 758.189092] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.194947] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 758.195154] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 758.196225] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dd852ea-1cbb-4188-8b39-e808afd80417 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.213926] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293533, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.217782] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 758.217782] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523e802d-5c8b-9f52-5cb9-d4336f6814b3" [ 758.217782] env[62814]: _type = "Task" [ 758.217782] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.226106] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523e802d-5c8b-9f52-5cb9-d4336f6814b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.227774] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53fe2a4-3c41-4f66-8ded-f04402fe843e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.235529] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae9ae26b-5e3d-48ac-88db-dda22b18851a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.240640] env[62814]: DEBUG nova.network.neutron [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Successfully updated port: 02f411b5-731e-45d9-a7dc-9aca642cd9aa {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 758.271042] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26181fa3-a433-4a46-9b9d-e3cef9b547f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.280141] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc89260e-e3cc-4953-92a9-0651c7bcee78 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.293646] env[62814]: DEBUG nova.compute.provider_tree [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 758.432395] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquiring lock "369cd937-4c18-4068-ae59-70a1d585094b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.432858] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Lock "369cd937-4c18-4068-ae59-70a1d585094b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.433172] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquiring lock "369cd937-4c18-4068-ae59-70a1d585094b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.433288] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Lock "369cd937-4c18-4068-ae59-70a1d585094b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.433496] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Lock "369cd937-4c18-4068-ae59-70a1d585094b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 758.436251] env[62814]: INFO nova.compute.manager [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Terminating instance [ 758.501428] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293532, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.654939] env[62814]: DEBUG oslo_concurrency.lockutils [req-33a1bc8d-2cba-4fbd-88b9-440600cd9f98 req-1819cda6-4a00-4ab5-a86d-8b607e2192b6 service nova] Releasing lock "refresh_cache-369cd937-4c18-4068-ae59-70a1d585094b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.661572] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 758.661970] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-af70c73d-c524-406c-bd1f-effcb2fa4441 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.670529] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 758.670529] env[62814]: value = "task-4293534" [ 758.670529] env[62814]: _type = "Task" [ 758.670529] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.679941] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293534, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.699559] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293533, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.732089] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523e802d-5c8b-9f52-5cb9-d4336f6814b3, 'name': SearchDatastore_Task, 'duration_secs': 0.029061} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.732869] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8efc29b6-c935-44e0-beb4-fc09c1f52f27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.740645] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 758.740645] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5296ebf1-2427-3db0-882c-25d4edccf82d" [ 758.740645] env[62814]: _type = "Task" [ 758.740645] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.744249] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquiring lock "refresh_cache-1a16bd56-0992-4bec-bd3f-2836dfe68579" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.744530] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquired lock "refresh_cache-1a16bd56-0992-4bec-bd3f-2836dfe68579" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.744705] env[62814]: DEBUG nova.network.neutron [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.753101] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5296ebf1-2427-3db0-882c-25d4edccf82d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.831457] env[62814]: DEBUG nova.scheduler.client.report [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 62 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 758.831857] env[62814]: DEBUG nova.compute.provider_tree [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 62 to 63 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 758.832211] env[62814]: DEBUG nova.compute.provider_tree [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 758.940328] env[62814]: DEBUG nova.compute.manager [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 758.940598] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 758.941631] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696ea203-53be-4ee7-a11a-a2d161ed0e28 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.949281] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 758.949436] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-748300c9-8f2e-4cd4-a310-d7c433fa593e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.955198] env[62814]: DEBUG oslo_vmware.api [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for the task: (returnval){ [ 758.955198] env[62814]: value = "task-4293535" [ 758.955198] env[62814]: _type = "Task" [ 758.955198] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.963091] env[62814]: DEBUG oslo_vmware.api [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293535, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.999781] env[62814]: DEBUG oslo_vmware.api [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293532, 'name': PowerOnVM_Task, 'duration_secs': 0.583364} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.000161] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 759.000445] env[62814]: INFO nova.compute.manager [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Took 9.23 seconds to spawn the instance on the hypervisor. [ 759.000648] env[62814]: DEBUG nova.compute.manager [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 759.001480] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24dd01f-0b14-40d9-94ae-2b17798fab45 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.180962] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293534, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.199145] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293533, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588127} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.199502] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 961dedc6-5ddd-4620-bfa7-54cf0c520af3/961dedc6-5ddd-4620-bfa7-54cf0c520af3.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 759.199763] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 759.200066] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-685f826e-09f7-40de-9548-8fba1c932a3c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.206974] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 759.206974] env[62814]: value = "task-4293536" [ 759.206974] env[62814]: _type = "Task" [ 759.206974] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.215688] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293536, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.252765] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5296ebf1-2427-3db0-882c-25d4edccf82d, 'name': SearchDatastore_Task, 'duration_secs': 0.021096} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.253101] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.253382] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 56f9e03c-1157-442f-8add-156627a6ec1e/56f9e03c-1157-442f-8add-156627a6ec1e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 759.254031] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e850c732-de01-446a-8312-526b2d782edd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.260087] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 759.260087] env[62814]: value = "task-4293537" [ 759.260087] env[62814]: _type = "Task" [ 759.260087] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.268861] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293537, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.298819] env[62814]: DEBUG nova.network.neutron [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.337314] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.024s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.337997] env[62814]: DEBUG nova.compute.manager [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 759.340900] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.338s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.341092] env[62814]: DEBUG nova.objects.instance [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lazy-loading 'resources' on Instance uuid 004fd137-4902-4313-a6f7-6c83cd76743d {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 759.464778] env[62814]: DEBUG oslo_vmware.api [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293535, 'name': PowerOffVM_Task, 'duration_secs': 0.20907} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.465068] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 759.465243] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 759.465492] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45e1f6c3-6bd1-42fd-bfc8-439b98cf5b2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.467983] env[62814]: DEBUG nova.network.neutron [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Updating instance_info_cache with network_info: [{"id": "02f411b5-731e-45d9-a7dc-9aca642cd9aa", "address": "fa:16:3e:55:12:ae", "network": {"id": "3cbd3440-c630-4c57-a5a9-d109b77a6064", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1749196117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8c68a04eda4221925a5397578c0280", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8ee8640-3787-4c27-9581-962ddb2be7e5", "external-id": "nsx-vlan-transportzone-224", "segmentation_id": 224, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02f411b5-73", "ovs_interfaceid": "02f411b5-731e-45d9-a7dc-9aca642cd9aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.519142] env[62814]: INFO nova.compute.manager [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Took 43.20 seconds to build instance. [ 759.533336] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 759.533609] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 759.533817] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Deleting the datastore file [datastore2] 369cd937-4c18-4068-ae59-70a1d585094b {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 759.534108] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fb93dfc-d4c4-4c14-934c-b95b69a13ba9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.543376] env[62814]: DEBUG oslo_vmware.api [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for the task: (returnval){ [ 759.543376] env[62814]: value = "task-4293539" [ 759.543376] env[62814]: _type = "Task" [ 759.543376] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.551720] env[62814]: DEBUG oslo_vmware.api [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293539, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.681785] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293534, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.717733] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293536, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066176} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.718092] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 759.719499] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b2ad4c-3bd6-4080-88b3-8a466d0b6e6a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.750025] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 961dedc6-5ddd-4620-bfa7-54cf0c520af3/961dedc6-5ddd-4620-bfa7-54cf0c520af3.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 759.750403] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cabb782-d90a-4d2d-8ec0-1ef23140d33d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.777862] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293537, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.779460] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 759.779460] env[62814]: value = "task-4293540" [ 759.779460] env[62814]: _type = "Task" [ 759.779460] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.789611] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293540, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.844682] env[62814]: DEBUG nova.compute.utils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 759.850262] env[62814]: DEBUG nova.compute.manager [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 759.850578] env[62814]: DEBUG nova.network.neutron [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 759.916416] env[62814]: DEBUG nova.policy [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3918eee433b4e48be96e92428ae500f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fccf636945024a0aa2c12c048e2e6a67', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 759.971483] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Releasing lock "refresh_cache-1a16bd56-0992-4bec-bd3f-2836dfe68579" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.971618] env[62814]: DEBUG nova.compute.manager [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Instance network_info: |[{"id": "02f411b5-731e-45d9-a7dc-9aca642cd9aa", "address": "fa:16:3e:55:12:ae", "network": {"id": "3cbd3440-c630-4c57-a5a9-d109b77a6064", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1749196117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8c68a04eda4221925a5397578c0280", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8ee8640-3787-4c27-9581-962ddb2be7e5", "external-id": "nsx-vlan-transportzone-224", "segmentation_id": 224, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02f411b5-73", "ovs_interfaceid": "02f411b5-731e-45d9-a7dc-9aca642cd9aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 759.972140] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:12:ae', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c8ee8640-3787-4c27-9581-962ddb2be7e5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '02f411b5-731e-45d9-a7dc-9aca642cd9aa', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 759.982279] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Creating folder: Project (0e8c68a04eda4221925a5397578c0280). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 759.982673] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-859de6e6-a41e-4d9e-a3de-07f556f81d94 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.002590] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Created folder: Project (0e8c68a04eda4221925a5397578c0280) in parent group-v845547. [ 760.002814] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Creating folder: Instances. Parent ref: group-v845648. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 760.003085] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e8df01e8-fb8e-455e-b99b-97f7d11223a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.015810] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Created folder: Instances in parent group-v845648. [ 760.016079] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 760.016303] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 760.016517] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6ac5b62-4c2a-4f1c-b80a-6b7a3e1b8b8e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.034860] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8523b3b8-2235-4cf8-a4b5-400375dbcf1b tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.408s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 760.043351] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 760.043351] env[62814]: value = "task-4293543" [ 760.043351] env[62814]: _type = "Task" [ 760.043351] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.059152] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293543, 'name': CreateVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.062485] env[62814]: DEBUG oslo_vmware.api [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293539, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.100466] env[62814]: DEBUG nova.compute.manager [req-64d4ea19-b9d0-460e-8c29-f36e69b26129 req-2de8ae14-f70d-4ad3-8891-b57c02e1628c service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Received event network-changed-02f411b5-731e-45d9-a7dc-9aca642cd9aa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 760.100706] env[62814]: DEBUG nova.compute.manager [req-64d4ea19-b9d0-460e-8c29-f36e69b26129 req-2de8ae14-f70d-4ad3-8891-b57c02e1628c service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Refreshing instance network info cache due to event network-changed-02f411b5-731e-45d9-a7dc-9aca642cd9aa. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 760.100925] env[62814]: DEBUG oslo_concurrency.lockutils [req-64d4ea19-b9d0-460e-8c29-f36e69b26129 req-2de8ae14-f70d-4ad3-8891-b57c02e1628c service nova] Acquiring lock "refresh_cache-1a16bd56-0992-4bec-bd3f-2836dfe68579" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.101072] env[62814]: DEBUG oslo_concurrency.lockutils [req-64d4ea19-b9d0-460e-8c29-f36e69b26129 req-2de8ae14-f70d-4ad3-8891-b57c02e1628c service nova] Acquired lock "refresh_cache-1a16bd56-0992-4bec-bd3f-2836dfe68579" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 760.101231] env[62814]: DEBUG nova.network.neutron [req-64d4ea19-b9d0-460e-8c29-f36e69b26129 req-2de8ae14-f70d-4ad3-8891-b57c02e1628c service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Refreshing network info cache for port 02f411b5-731e-45d9-a7dc-9aca642cd9aa {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 760.185269] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293534, 'name': CreateSnapshot_Task, 'duration_secs': 1.468544} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.185556] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 760.186335] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c9ec5a-6d8e-4a8f-9bda-6ffb6dc027b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.280797] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293537, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.84447} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.286281] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 56f9e03c-1157-442f-8add-156627a6ec1e/56f9e03c-1157-442f-8add-156627a6ec1e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 760.286541] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 760.287027] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-add0ba18-2bf0-4329-9fb5-4338a83e8e16 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.289641] env[62814]: DEBUG nova.network.neutron [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Successfully created port: c108416a-2aa8-4882-a9a2-d0cbfbb4db61 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.301245] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293540, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.301561] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 760.301561] env[62814]: value = "task-4293544" [ 760.301561] env[62814]: _type = "Task" [ 760.301561] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.314123] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293544, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.352196] env[62814]: DEBUG nova.compute.manager [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 760.433631] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f42daa3-d0ae-4b61-98a5-9df174c5f57a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.442412] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3cd197-cb4a-4e82-bc26-e6f58ff463a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.477171] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c732e5-29cd-4490-8fe8-1c9ca54d2e90 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.485087] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b6f0d8-3bf3-4a17-9b0e-b486945ab7ad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.500705] env[62814]: DEBUG nova.compute.provider_tree [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.537797] env[62814]: DEBUG nova.compute.manager [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 760.558294] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293543, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.563630] env[62814]: DEBUG oslo_vmware.api [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Task: {'id': task-4293539, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.557407} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.564034] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 760.564121] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 760.564588] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 760.564588] env[62814]: INFO nova.compute.manager [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Took 1.62 seconds to destroy the instance on the hypervisor. [ 760.564693] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 760.565050] env[62814]: DEBUG nova.compute.manager [-] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 760.565050] env[62814]: DEBUG nova.network.neutron [-] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.638920] env[62814]: DEBUG nova.compute.manager [req-b26ea01a-dfad-404d-9ffb-42bb5be79fae req-d62d52be-2cdd-43a9-83e7-470af3f3eb5a service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Received event network-changed-08bab1d2-6296-46f7-baf6-4344d1bbb0ef {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 760.642649] env[62814]: DEBUG nova.compute.manager [req-b26ea01a-dfad-404d-9ffb-42bb5be79fae req-d62d52be-2cdd-43a9-83e7-470af3f3eb5a service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Refreshing instance network info cache due to event network-changed-08bab1d2-6296-46f7-baf6-4344d1bbb0ef. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 760.642649] env[62814]: DEBUG oslo_concurrency.lockutils [req-b26ea01a-dfad-404d-9ffb-42bb5be79fae req-d62d52be-2cdd-43a9-83e7-470af3f3eb5a service nova] Acquiring lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.642649] env[62814]: DEBUG oslo_concurrency.lockutils [req-b26ea01a-dfad-404d-9ffb-42bb5be79fae req-d62d52be-2cdd-43a9-83e7-470af3f3eb5a service nova] Acquired lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 760.642649] env[62814]: DEBUG nova.network.neutron [req-b26ea01a-dfad-404d-9ffb-42bb5be79fae req-d62d52be-2cdd-43a9-83e7-470af3f3eb5a service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Refreshing network info cache for port 08bab1d2-6296-46f7-baf6-4344d1bbb0ef {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 760.714080] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 760.714080] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e94e7182-55e5-4f72-b0d2-08b1cfb7402f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.723682] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 760.723682] env[62814]: value = "task-4293545" [ 760.723682] env[62814]: _type = "Task" [ 760.723682] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.732946] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293545, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.791959] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293540, 'name': ReconfigVM_Task, 'duration_secs': 0.709996} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.792457] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 961dedc6-5ddd-4620-bfa7-54cf0c520af3/961dedc6-5ddd-4620-bfa7-54cf0c520af3.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 760.793349] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe40791f-14bc-4a90-825e-724dfd539896 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.800284] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 760.800284] env[62814]: value = "task-4293546" [ 760.800284] env[62814]: _type = "Task" [ 760.800284] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.811805] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293546, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.817052] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293544, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070767} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.819997] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 760.820832] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a091c20-b0ff-41d4-ab5b-1ddf1fe9843f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.845111] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] 56f9e03c-1157-442f-8add-156627a6ec1e/56f9e03c-1157-442f-8add-156627a6ec1e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 760.845483] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c952c7e-749d-4110-85ba-d85b2fc4aa48 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.874927] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 760.874927] env[62814]: value = "task-4293547" [ 760.874927] env[62814]: _type = "Task" [ 760.874927] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.884964] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293547, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.893565] env[62814]: DEBUG nova.network.neutron [req-64d4ea19-b9d0-460e-8c29-f36e69b26129 req-2de8ae14-f70d-4ad3-8891-b57c02e1628c service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Updated VIF entry in instance network info cache for port 02f411b5-731e-45d9-a7dc-9aca642cd9aa. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 760.893932] env[62814]: DEBUG nova.network.neutron [req-64d4ea19-b9d0-460e-8c29-f36e69b26129 req-2de8ae14-f70d-4ad3-8891-b57c02e1628c service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Updating instance_info_cache with network_info: [{"id": "02f411b5-731e-45d9-a7dc-9aca642cd9aa", "address": "fa:16:3e:55:12:ae", "network": {"id": "3cbd3440-c630-4c57-a5a9-d109b77a6064", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1749196117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8c68a04eda4221925a5397578c0280", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8ee8640-3787-4c27-9581-962ddb2be7e5", "external-id": "nsx-vlan-transportzone-224", "segmentation_id": 224, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02f411b5-73", "ovs_interfaceid": "02f411b5-731e-45d9-a7dc-9aca642cd9aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.004555] env[62814]: DEBUG nova.scheduler.client.report [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 761.055873] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293543, 'name': CreateVM_Task, 'duration_secs': 0.751758} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.056148] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 761.056952] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.057177] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 761.057502] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 761.058265] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3736762b-876f-40f8-a1d6-153d90480688 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.064185] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 761.064185] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523c8d7d-393a-fc5b-83a5-933a52f2e46a" [ 761.064185] env[62814]: _type = "Task" [ 761.064185] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.065192] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 761.073471] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523c8d7d-393a-fc5b-83a5-933a52f2e46a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.234836] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293545, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.311165] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293546, 'name': Rename_Task, 'duration_secs': 0.188065} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.311448] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 761.311739] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-941a620d-6ed9-4395-a24b-385a7715c3c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.321895] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 761.321895] env[62814]: value = "task-4293548" [ 761.321895] env[62814]: _type = "Task" [ 761.321895] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.331416] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293548, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.372171] env[62814]: DEBUG nova.compute.manager [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 761.387244] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293547, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.399452] env[62814]: DEBUG oslo_concurrency.lockutils [req-64d4ea19-b9d0-460e-8c29-f36e69b26129 req-2de8ae14-f70d-4ad3-8891-b57c02e1628c service nova] Releasing lock "refresh_cache-1a16bd56-0992-4bec-bd3f-2836dfe68579" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 761.420124] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 761.420306] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.420542] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 761.420697] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.420822] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 761.421052] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 761.421452] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 761.421734] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 761.422064] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 761.422295] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 761.422539] env[62814]: DEBUG nova.virt.hardware [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 761.423942] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edbcb461-46c7-4a13-bbce-199a4fee2a7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.433114] env[62814]: DEBUG nova.network.neutron [-] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.439200] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235b74b4-c616-41f3-87d0-2449713fef98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.483925] env[62814]: DEBUG nova.network.neutron [req-b26ea01a-dfad-404d-9ffb-42bb5be79fae req-d62d52be-2cdd-43a9-83e7-470af3f3eb5a service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updated VIF entry in instance network info cache for port 08bab1d2-6296-46f7-baf6-4344d1bbb0ef. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 761.484547] env[62814]: DEBUG nova.network.neutron [req-b26ea01a-dfad-404d-9ffb-42bb5be79fae req-d62d52be-2cdd-43a9-83e7-470af3f3eb5a service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance_info_cache with network_info: [{"id": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "address": "fa:16:3e:db:ee:dc", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08bab1d2-62", "ovs_interfaceid": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.511089] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.170s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 761.514673] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.978s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 761.515113] env[62814]: DEBUG nova.objects.instance [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lazy-loading 'resources' on Instance uuid e6af4651-9f3a-4ce0-add8-06f1cfef255f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 761.540631] env[62814]: INFO nova.scheduler.client.report [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Deleted allocations for instance 004fd137-4902-4313-a6f7-6c83cd76743d [ 761.576547] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523c8d7d-393a-fc5b-83a5-933a52f2e46a, 'name': SearchDatastore_Task, 'duration_secs': 0.032234} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.576933] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 761.577314] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 761.577734] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.577897] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 761.578196] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 761.579306] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f051f864-9da2-4271-9c80-dcfa1f01a9bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.589234] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 761.589580] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 761.591018] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3302fa4-20e6-4d5a-96d5-d4e9b7e678d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.598505] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 761.598505] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b5967c-1c8d-8036-c456-8685f0d00321" [ 761.598505] env[62814]: _type = "Task" [ 761.598505] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.613812] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b5967c-1c8d-8036-c456-8685f0d00321, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.739400] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293545, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.833135] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293548, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.886050] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293547, 'name': ReconfigVM_Task, 'duration_secs': 0.667977} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.886349] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Reconfigured VM instance instance-00000024 to attach disk [datastore2] 56f9e03c-1157-442f-8add-156627a6ec1e/56f9e03c-1157-442f-8add-156627a6ec1e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.887043] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6968b7cb-e977-4b17-be14-783e07551e56 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.895169] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 761.895169] env[62814]: value = "task-4293549" [ 761.895169] env[62814]: _type = "Task" [ 761.895169] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.907972] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293549, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.942958] env[62814]: INFO nova.compute.manager [-] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Took 1.38 seconds to deallocate network for instance. [ 761.988601] env[62814]: DEBUG oslo_concurrency.lockutils [req-b26ea01a-dfad-404d-9ffb-42bb5be79fae req-d62d52be-2cdd-43a9-83e7-470af3f3eb5a service nova] Releasing lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 762.039475] env[62814]: DEBUG nova.network.neutron [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Successfully updated port: c108416a-2aa8-4882-a9a2-d0cbfbb4db61 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 762.056219] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5f342372-147b-4b14-a0d5-36d15853d070 tempest-ServerAddressesNegativeTestJSON-2036518242 tempest-ServerAddressesNegativeTestJSON-2036518242-project-member] Lock "004fd137-4902-4313-a6f7-6c83cd76743d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.574s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.109634] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b5967c-1c8d-8036-c456-8685f0d00321, 'name': SearchDatastore_Task, 'duration_secs': 0.011489} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.113555] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c4fe1bd-68da-4c89-bc32-e4a410ee7354 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.119806] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 762.119806] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dbc9e4-362b-9000-31bd-38ffa92bd134" [ 762.119806] env[62814]: _type = "Task" [ 762.119806] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.129259] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dbc9e4-362b-9000-31bd-38ffa92bd134, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.167344] env[62814]: DEBUG nova.compute.manager [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Received event network-vif-deleted-efe64bba-a178-4079-b256-51b9e74293c3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 762.167344] env[62814]: DEBUG nova.compute.manager [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Received event network-vif-plugged-c108416a-2aa8-4882-a9a2-d0cbfbb4db61 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 762.167344] env[62814]: DEBUG oslo_concurrency.lockutils [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] Acquiring lock "9677b03f-7138-47b7-b1e8-f3714d11e550-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.167344] env[62814]: DEBUG oslo_concurrency.lockutils [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] Lock "9677b03f-7138-47b7-b1e8-f3714d11e550-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 762.167344] env[62814]: DEBUG oslo_concurrency.lockutils [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] Lock "9677b03f-7138-47b7-b1e8-f3714d11e550-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.167344] env[62814]: DEBUG nova.compute.manager [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] No waiting events found dispatching network-vif-plugged-c108416a-2aa8-4882-a9a2-d0cbfbb4db61 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 762.167344] env[62814]: WARNING nova.compute.manager [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Received unexpected event network-vif-plugged-c108416a-2aa8-4882-a9a2-d0cbfbb4db61 for instance with vm_state building and task_state spawning. [ 762.167344] env[62814]: DEBUG nova.compute.manager [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Received event network-changed-c108416a-2aa8-4882-a9a2-d0cbfbb4db61 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 762.167597] env[62814]: DEBUG nova.compute.manager [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Refreshing instance network info cache due to event network-changed-c108416a-2aa8-4882-a9a2-d0cbfbb4db61. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 762.167597] env[62814]: DEBUG oslo_concurrency.lockutils [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] Acquiring lock "refresh_cache-9677b03f-7138-47b7-b1e8-f3714d11e550" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.167718] env[62814]: DEBUG oslo_concurrency.lockutils [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] Acquired lock "refresh_cache-9677b03f-7138-47b7-b1e8-f3714d11e550" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 762.167876] env[62814]: DEBUG nova.network.neutron [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Refreshing network info cache for port c108416a-2aa8-4882-a9a2-d0cbfbb4db61 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 762.238614] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293545, 'name': CloneVM_Task} progress is 95%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.333836] env[62814]: DEBUG oslo_vmware.api [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293548, 'name': PowerOnVM_Task, 'duration_secs': 0.713525} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.334183] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 762.334685] env[62814]: INFO nova.compute.manager [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Took 15.35 seconds to spawn the instance on the hypervisor. [ 762.334685] env[62814]: DEBUG nova.compute.manager [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 762.335484] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754e01ae-ab20-410e-96dc-217cee9177f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.406282] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293549, 'name': Rename_Task, 'duration_secs': 0.282029} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.409234] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 762.409777] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31f8b729-d4fd-4bad-8fd2-71a61b0452a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.417340] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 762.417340] env[62814]: value = "task-4293550" [ 762.417340] env[62814]: _type = "Task" [ 762.417340] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.431253] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293550, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.450847] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.542373] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquiring lock "refresh_cache-9677b03f-7138-47b7-b1e8-f3714d11e550" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.606904] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a2c647-4014-42a2-9012-8f27ce3671e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.616233] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77071891-9d3f-485d-8adc-454d0050968b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.630846] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dbc9e4-362b-9000-31bd-38ffa92bd134, 'name': SearchDatastore_Task, 'duration_secs': 0.011489} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.657047] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 762.657047] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1a16bd56-0992-4bec-bd3f-2836dfe68579/1a16bd56-0992-4bec-bd3f-2836dfe68579.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 762.657047] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-229f0d9b-5310-4b2a-9062-f03a42f2017b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.660009] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf8fb683-38fd-457d-a6c4-f758ae89fca3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.667262] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 762.667262] env[62814]: value = "task-4293551" [ 762.667262] env[62814]: _type = "Task" [ 762.667262] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.672261] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f55c6f-f436-4590-82b1-67b396f30aab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.685141] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293551, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.691069] env[62814]: DEBUG nova.compute.provider_tree [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.724627] env[62814]: DEBUG nova.network.neutron [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.739968] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293545, 'name': CloneVM_Task, 'duration_secs': 1.68655} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.740308] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Created linked-clone VM from snapshot [ 762.741056] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb7c9a1-03d2-43de-9029-1a0829893b1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.748540] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Uploading image f5e1a700-0673-4eca-adf9-4520693e9282 {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 762.780799] env[62814]: DEBUG oslo_vmware.rw_handles [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 762.780799] env[62814]: value = "vm-845652" [ 762.780799] env[62814]: _type = "VirtualMachine" [ 762.780799] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 762.781214] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-711c6982-4518-4a24-8d7b-aafe2fb77a21 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.789179] env[62814]: DEBUG oslo_vmware.rw_handles [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lease: (returnval){ [ 762.789179] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b4ab0-7ef7-9b96-4a44-876ee2c7c97c" [ 762.789179] env[62814]: _type = "HttpNfcLease" [ 762.789179] env[62814]: } obtained for exporting VM: (result){ [ 762.789179] env[62814]: value = "vm-845652" [ 762.789179] env[62814]: _type = "VirtualMachine" [ 762.789179] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 762.789525] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the lease: (returnval){ [ 762.789525] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b4ab0-7ef7-9b96-4a44-876ee2c7c97c" [ 762.789525] env[62814]: _type = "HttpNfcLease" [ 762.789525] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 762.797987] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 762.797987] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b4ab0-7ef7-9b96-4a44-876ee2c7c97c" [ 762.797987] env[62814]: _type = "HttpNfcLease" [ 762.797987] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 762.857585] env[62814]: INFO nova.compute.manager [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Took 46.96 seconds to build instance. [ 762.870213] env[62814]: DEBUG nova.network.neutron [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.929924] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293550, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.187836] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293551, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.198035] env[62814]: DEBUG nova.scheduler.client.report [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 763.300041] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 763.300041] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b4ab0-7ef7-9b96-4a44-876ee2c7c97c" [ 763.300041] env[62814]: _type = "HttpNfcLease" [ 763.300041] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 763.300521] env[62814]: DEBUG oslo_vmware.rw_handles [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 763.300521] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b4ab0-7ef7-9b96-4a44-876ee2c7c97c" [ 763.300521] env[62814]: _type = "HttpNfcLease" [ 763.300521] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 763.302318] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c4cb07-ca1a-4c24-a1c1-b753cc59238f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.311196] env[62814]: DEBUG oslo_vmware.rw_handles [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a257d-7a25-e314-6f4b-e79eb6c30a9c/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 763.311373] env[62814]: DEBUG oslo_vmware.rw_handles [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a257d-7a25-e314-6f4b-e79eb6c30a9c/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 763.366750] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7c45a773-840e-4958-ad2c-871103e446ab tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 78.890s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.372438] env[62814]: DEBUG oslo_concurrency.lockutils [req-6e47c64d-1b6d-4812-9eef-316502d0a2d2 req-bd300cca-9634-40c4-bb8d-64c8b1016f5f service nova] Releasing lock "refresh_cache-9677b03f-7138-47b7-b1e8-f3714d11e550" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 763.372877] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquired lock "refresh_cache-9677b03f-7138-47b7-b1e8-f3714d11e550" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.373075] env[62814]: DEBUG nova.network.neutron [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.429171] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293550, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.556040] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-71646f9d-78ce-4d92-8463-d4b51044f325 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.685169] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293551, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.648278} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.685503] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1a16bd56-0992-4bec-bd3f-2836dfe68579/1a16bd56-0992-4bec-bd3f-2836dfe68579.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 763.685737] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 763.685986] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a0c1887-67bc-4b2a-9a05-0da8b6693d47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.697025] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 763.697025] env[62814]: value = "task-4293553" [ 763.697025] env[62814]: _type = "Task" [ 763.697025] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.708952] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.194s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.714153] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293553, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.714153] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.755s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.714153] env[62814]: DEBUG nova.objects.instance [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lazy-loading 'resources' on Instance uuid be02bc93-1278-4b3d-afa3-270c84585d1d {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 763.756465] env[62814]: INFO nova.scheduler.client.report [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Deleted allocations for instance e6af4651-9f3a-4ce0-add8-06f1cfef255f [ 763.872614] env[62814]: DEBUG nova.compute.manager [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 763.926394] env[62814]: DEBUG nova.network.neutron [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.938442] env[62814]: DEBUG oslo_vmware.api [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293550, 'name': PowerOnVM_Task, 'duration_secs': 1.016513} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.938755] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 763.938962] env[62814]: INFO nova.compute.manager [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Took 11.44 seconds to spawn the instance on the hypervisor. [ 763.939190] env[62814]: DEBUG nova.compute.manager [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 763.940069] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35839232-853c-44e9-9cca-c6ff3b455d75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.969131] env[62814]: DEBUG oslo_concurrency.lockutils [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 763.969636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.969731] env[62814]: DEBUG oslo_concurrency.lockutils [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 763.969852] env[62814]: DEBUG oslo_concurrency.lockutils [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.970143] env[62814]: DEBUG oslo_concurrency.lockutils [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.976526] env[62814]: INFO nova.compute.manager [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Terminating instance [ 764.142087] env[62814]: DEBUG nova.network.neutron [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Updating instance_info_cache with network_info: [{"id": "c108416a-2aa8-4882-a9a2-d0cbfbb4db61", "address": "fa:16:3e:c0:73:54", "network": {"id": "e3298aaa-0875-4267-8cb7-8656c9d2d087", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-260246757-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fccf636945024a0aa2c12c048e2e6a67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc108416a-2a", "ovs_interfaceid": "c108416a-2aa8-4882-a9a2-d0cbfbb4db61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.206355] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293553, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.12204} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.206630] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 764.207444] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a305018a-6e12-4163-a4fe-bd526786aacc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.238263] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Reconfiguring VM instance instance-00000025 to attach disk [datastore2] 1a16bd56-0992-4bec-bd3f-2836dfe68579/1a16bd56-0992-4bec-bd3f-2836dfe68579.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 764.239496] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76efdee7-2c48-4dec-817d-f9d1173455f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.263574] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 764.263574] env[62814]: value = "task-4293554" [ 764.263574] env[62814]: _type = "Task" [ 764.263574] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.270447] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cf9e9640-5514-4ae2-957f-5d56ff3606e8 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "e6af4651-9f3a-4ce0-add8-06f1cfef255f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.232s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.283860] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.407548] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 764.462510] env[62814]: INFO nova.compute.manager [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Took 41.67 seconds to build instance. [ 764.482375] env[62814]: DEBUG nova.compute.manager [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 764.482712] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 764.486761] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f277c139-1ffa-4bd7-aa42-ffa47c7f8908 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.495072] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 764.495388] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-431c5f1b-b02c-49af-83db-590d4b04d8e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.503838] env[62814]: DEBUG oslo_vmware.api [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 764.503838] env[62814]: value = "task-4293555" [ 764.503838] env[62814]: _type = "Task" [ 764.503838] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.512631] env[62814]: DEBUG oslo_vmware.api [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293555, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.647035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Releasing lock "refresh_cache-9677b03f-7138-47b7-b1e8-f3714d11e550" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 764.647035] env[62814]: DEBUG nova.compute.manager [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Instance network_info: |[{"id": "c108416a-2aa8-4882-a9a2-d0cbfbb4db61", "address": "fa:16:3e:c0:73:54", "network": {"id": "e3298aaa-0875-4267-8cb7-8656c9d2d087", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-260246757-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fccf636945024a0aa2c12c048e2e6a67", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4adc8ed0-d11a-4510-9be0-b27c0da3a903", "external-id": "nsx-vlan-transportzone-844", "segmentation_id": 844, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc108416a-2a", "ovs_interfaceid": "c108416a-2aa8-4882-a9a2-d0cbfbb4db61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 764.647391] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c0:73:54', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4adc8ed0-d11a-4510-9be0-b27c0da3a903', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c108416a-2aa8-4882-a9a2-d0cbfbb4db61', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 764.660459] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Creating folder: Project (fccf636945024a0aa2c12c048e2e6a67). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 764.661035] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-51e8af08-bd64-467b-98a2-023558923272 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.677973] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Created folder: Project (fccf636945024a0aa2c12c048e2e6a67) in parent group-v845547. [ 764.678262] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Creating folder: Instances. Parent ref: group-v845653. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 764.681338] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3e05c997-3305-4643-be26-f91d27992dd3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.696679] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Created folder: Instances in parent group-v845653. [ 764.696999] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 764.698135] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 764.698967] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c5c84b11-6171-43c4-b570-7eedc82cc40c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.728208] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 764.728208] env[62814]: value = "task-4293558" [ 764.728208] env[62814]: _type = "Task" [ 764.728208] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.740652] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293558, 'name': CreateVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.789669] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293554, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.922250] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affdc123-5323-4741-b28b-b4cda996020e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.932243] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf21f6a1-adc1-4be1-b426-0ace1444e9a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.974722] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7a1d8ff-0d55-4104-b11e-4d9732a67e91 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "56f9e03c-1157-442f-8add-156627a6ec1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.235s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.977591] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba54182a-c772-441a-ab49-220d560acce9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.989764] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e05078c-92be-4423-a1c9-832177de774a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.011027] env[62814]: DEBUG nova.compute.provider_tree [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.025676] env[62814]: DEBUG oslo_vmware.api [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293555, 'name': PowerOffVM_Task, 'duration_secs': 0.200665} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.025971] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 765.026160] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 765.026420] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e4c3bb6-ebc5-4281-9501-d1736d424340 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.150621] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 765.150959] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 765.151895] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Deleting the datastore file [datastore2] 961dedc6-5ddd-4620-bfa7-54cf0c520af3 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 765.151895] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-099db46e-5e62-4064-9cc6-6fa8ed938500 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.161834] env[62814]: DEBUG oslo_vmware.api [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 765.161834] env[62814]: value = "task-4293560" [ 765.161834] env[62814]: _type = "Task" [ 765.161834] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.170780] env[62814]: DEBUG oslo_vmware.api [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293560, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.200538] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "56f9e03c-1157-442f-8add-156627a6ec1e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 765.201080] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "56f9e03c-1157-442f-8add-156627a6ec1e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 765.201080] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "56f9e03c-1157-442f-8add-156627a6ec1e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 765.201214] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "56f9e03c-1157-442f-8add-156627a6ec1e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 765.201369] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "56f9e03c-1157-442f-8add-156627a6ec1e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.203704] env[62814]: INFO nova.compute.manager [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Terminating instance [ 765.240377] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293558, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.274319] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293554, 'name': ReconfigVM_Task, 'duration_secs': 0.935599} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.274663] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Reconfigured VM instance instance-00000025 to attach disk [datastore2] 1a16bd56-0992-4bec-bd3f-2836dfe68579/1a16bd56-0992-4bec-bd3f-2836dfe68579.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 765.275333] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65b3c1ae-3830-4f5c-9943-e3e9b65c84ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.282386] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 765.282386] env[62814]: value = "task-4293561" [ 765.282386] env[62814]: _type = "Task" [ 765.282386] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.291255] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293561, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.484724] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 765.518913] env[62814]: DEBUG nova.scheduler.client.report [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 765.618202] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "59d21ef7-df97-49ac-9329-4c18df6dd087" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 765.618202] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 765.618202] env[62814]: DEBUG nova.compute.manager [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 765.618832] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc36a6a8-7619-4896-a575-c113978f53da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.629074] env[62814]: DEBUG nova.compute.manager [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 765.630234] env[62814]: DEBUG nova.objects.instance [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'flavor' on Instance uuid 59d21ef7-df97-49ac-9329-4c18df6dd087 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 765.670507] env[62814]: DEBUG oslo_vmware.api [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293560, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.395067} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.670794] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 765.670973] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 765.671408] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 765.671485] env[62814]: INFO nova.compute.manager [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Took 1.19 seconds to destroy the instance on the hypervisor. [ 765.671797] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 765.672019] env[62814]: DEBUG nova.compute.manager [-] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 765.672121] env[62814]: DEBUG nova.network.neutron [-] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.709096] env[62814]: DEBUG nova.compute.manager [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 765.709893] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 765.712330] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0ca3bb-a337-43ec-8a11-a4e889267f92 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.724209] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 765.724564] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c395259-b363-4f10-acbc-e569079e8f0d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.738025] env[62814]: DEBUG oslo_vmware.api [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 765.738025] env[62814]: value = "task-4293562" [ 765.738025] env[62814]: _type = "Task" [ 765.738025] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.744407] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293558, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.750457] env[62814]: DEBUG oslo_vmware.api [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293562, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.798267] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293561, 'name': Rename_Task, 'duration_secs': 0.277381} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.798696] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 765.799159] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49ac9faf-ab6c-463a-9b73-161bc5cf3690 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.806594] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 765.806594] env[62814]: value = "task-4293563" [ 765.806594] env[62814]: _type = "Task" [ 765.806594] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.815986] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293563, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.005450] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.026709] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.315s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.033771] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.731s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 766.035543] env[62814]: INFO nova.compute.claims [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.067706] env[62814]: INFO nova.scheduler.client.report [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Deleted allocations for instance be02bc93-1278-4b3d-afa3-270c84585d1d [ 766.239175] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293558, 'name': CreateVM_Task, 'duration_secs': 1.318413} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.242414] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 766.243104] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.243283] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.243599] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 766.244202] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5393428d-bbaf-4c4a-88b0-470b525ee668 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.252010] env[62814]: DEBUG oslo_vmware.api [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293562, 'name': PowerOffVM_Task, 'duration_secs': 0.196523} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.252714] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 766.252894] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 766.253184] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b3a31b5-0944-4334-9b20-70badf2694bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.256586] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 766.256586] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f74f1-5e8f-0825-7e90-7301c5ff47ca" [ 766.256586] env[62814]: _type = "Task" [ 766.256586] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.263836] env[62814]: DEBUG nova.compute.manager [req-8a8d74ec-71ac-416e-b0cc-404d45547100 req-824b7d3a-d190-4d0f-9758-b5be01c6cfe0 service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received event network-vif-deleted-92958ac5-6305-4ce3-aa70-0c6080099c48 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 766.264435] env[62814]: INFO nova.compute.manager [req-8a8d74ec-71ac-416e-b0cc-404d45547100 req-824b7d3a-d190-4d0f-9758-b5be01c6cfe0 service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Neutron deleted interface 92958ac5-6305-4ce3-aa70-0c6080099c48; detaching it from the instance and deleting it from the info cache [ 766.264505] env[62814]: DEBUG nova.network.neutron [req-8a8d74ec-71ac-416e-b0cc-404d45547100 req-824b7d3a-d190-4d0f-9758-b5be01c6cfe0 service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Updating instance_info_cache with network_info: [{"id": "ce7e8021-3ebf-4e8c-a0c2-618fa44c6957", "address": "fa:16:3e:2d:0b:38", "network": {"id": "4a0aa60f-1f50-4cc4-b64e-96e75cfc8d42", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1969588786", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.196", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ee1c76d-1a61-4546-85cb-d4bd3c1b35ef", "external-id": "nsx-vlan-transportzone-161", "segmentation_id": 161, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce7e8021-3e", "ovs_interfaceid": "ce7e8021-3ebf-4e8c-a0c2-618fa44c6957", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fbfe5ed2-1472-41c4-a028-38e81d37b3fe", "address": "fa:16:3e:1c:21:12", "network": {"id": "4fa67c69-1661-4203-a301-da311d1b4489", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-824717230", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.25", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50886eea-591a-452c-a27b-5f22cfc9df85", "external-id": "nsx-vlan-transportzone-578", "segmentation_id": 578, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfbfe5ed2-14", "ovs_interfaceid": "fbfe5ed2-1472-41c4-a028-38e81d37b3fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.270146] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f74f1-5e8f-0825-7e90-7301c5ff47ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.313965] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 766.314252] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 766.314444] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Deleting the datastore file [datastore2] 56f9e03c-1157-442f-8add-156627a6ec1e {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 766.315151] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae853567-6048-4873-abcd-fe5db9e148c2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.321015] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293563, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.325836] env[62814]: DEBUG oslo_vmware.api [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for the task: (returnval){ [ 766.325836] env[62814]: value = "task-4293565" [ 766.325836] env[62814]: _type = "Task" [ 766.325836] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.337858] env[62814]: DEBUG oslo_vmware.api [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293565, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.584654] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3b8d8688-5c08-480d-9c96-d549d3abcce4 tempest-ServersAdminNegativeTestJSON-1453842836 tempest-ServersAdminNegativeTestJSON-1453842836-project-member] Lock "be02bc93-1278-4b3d-afa3-270c84585d1d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.081s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.641021] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 766.641021] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-83cd4b57-6577-4c70-aaf7-5107bb8e547f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.650026] env[62814]: DEBUG oslo_vmware.api [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 766.650026] env[62814]: value = "task-4293566" [ 766.650026] env[62814]: _type = "Task" [ 766.650026] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.659532] env[62814]: DEBUG oslo_vmware.api [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293566, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.769273] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f74f1-5e8f-0825-7e90-7301c5ff47ca, 'name': SearchDatastore_Task, 'duration_secs': 0.013038} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.769666] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.769987] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 766.770908] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.770908] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 766.770908] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 766.773103] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b8d75ed-e2d1-4242-bb6e-91a3b6c7fc0e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.773796] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59edc990-7299-42a3-b2a9-79525aa1d4f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.787179] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061d7ba3-c5bb-4610-b15b-beaff325b3f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.802906] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 766.803334] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 766.804423] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe1bb3ab-3452-4d8b-8e40-3ce15a4ed2c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.814968] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 766.814968] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5235ae6a-c48b-64dc-4ba1-f38f511a1efe" [ 766.814968] env[62814]: _type = "Task" [ 766.814968] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.837559] env[62814]: DEBUG nova.compute.manager [req-8a8d74ec-71ac-416e-b0cc-404d45547100 req-824b7d3a-d190-4d0f-9758-b5be01c6cfe0 service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Detach interface failed, port_id=92958ac5-6305-4ce3-aa70-0c6080099c48, reason: Instance 961dedc6-5ddd-4620-bfa7-54cf0c520af3 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 766.856903] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293563, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.863491] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5235ae6a-c48b-64dc-4ba1-f38f511a1efe, 'name': SearchDatastore_Task, 'duration_secs': 0.019201} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.870445] env[62814]: DEBUG oslo_vmware.api [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Task: {'id': task-4293565, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.402861} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.870445] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5362bc0e-6e1c-4632-8fe5-7fa0a4590ebe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.871109] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 766.871174] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 766.871393] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 766.871748] env[62814]: INFO nova.compute.manager [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Took 1.16 seconds to destroy the instance on the hypervisor. [ 766.872053] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 766.874036] env[62814]: DEBUG nova.compute.manager [-] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 766.874130] env[62814]: DEBUG nova.network.neutron [-] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 766.885808] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 766.885808] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5299a9d6-1e5d-04df-0dc7-1aaa78456dba" [ 766.885808] env[62814]: _type = "Task" [ 766.885808] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.894234] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5299a9d6-1e5d-04df-0dc7-1aaa78456dba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.164058] env[62814]: DEBUG oslo_vmware.api [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293566, 'name': PowerOffVM_Task, 'duration_secs': 0.213402} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.164219] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 767.164762] env[62814]: DEBUG nova.compute.manager [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 767.165216] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d4a00e-5d42-496a-bb69-b4986d64ff0c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.322215] env[62814]: DEBUG oslo_vmware.api [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293563, 'name': PowerOnVM_Task, 'duration_secs': 1.401256} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.322858] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 767.323483] env[62814]: INFO nova.compute.manager [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Took 9.98 seconds to spawn the instance on the hypervisor. [ 767.323630] env[62814]: DEBUG nova.compute.manager [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 767.324495] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b80210-4d76-44b6-834d-fd5bf714fbe6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.380784] env[62814]: DEBUG nova.network.neutron [-] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.398175] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5299a9d6-1e5d-04df-0dc7-1aaa78456dba, 'name': SearchDatastore_Task, 'duration_secs': 0.012695} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.398175] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 767.398331] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 9677b03f-7138-47b7-b1e8-f3714d11e550/9677b03f-7138-47b7-b1e8-f3714d11e550.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 767.398804] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-45ce3f79-6d98-497c-895d-334f06457a9b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.410645] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 767.410645] env[62814]: value = "task-4293567" [ 767.410645] env[62814]: _type = "Task" [ 767.410645] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.421210] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.516021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "961f313b-b43f-4531-8a4b-0a39421d6a34" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.516021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "961f313b-b43f-4531-8a4b-0a39421d6a34" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.516021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "961f313b-b43f-4531-8a4b-0a39421d6a34-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.516021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "961f313b-b43f-4531-8a4b-0a39421d6a34-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.516021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "961f313b-b43f-4531-8a4b-0a39421d6a34-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.518815] env[62814]: INFO nova.compute.manager [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Terminating instance [ 767.639381] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "64a475e4-6713-408b-a63a-a43b5fed5ec8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.639381] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "64a475e4-6713-408b-a63a-a43b5fed5ec8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.639381] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "64a475e4-6713-408b-a63a-a43b5fed5ec8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.639381] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "64a475e4-6713-408b-a63a-a43b5fed5ec8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.641108] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "64a475e4-6713-408b-a63a-a43b5fed5ec8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.645991] env[62814]: INFO nova.compute.manager [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Terminating instance [ 767.692796] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8ffebbd9-3e15-4fde-ba51-fcc83909c09d tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.075s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.722013] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d0f5297-210c-46c6-b6b7-0587e7c4d9ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.731873] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654f1cb2-366a-49c8-aa04-04384f2b3ec9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.765832] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0efb981-c53a-4388-8440-ec35afe10dd8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.768791] env[62814]: DEBUG nova.network.neutron [-] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.776623] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e310f45-a553-4adb-85da-ed3ac007668d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.799984] env[62814]: DEBUG nova.compute.provider_tree [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.852763] env[62814]: INFO nova.compute.manager [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Took 38.73 seconds to build instance. [ 767.885162] env[62814]: INFO nova.compute.manager [-] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Took 2.21 seconds to deallocate network for instance. [ 767.925019] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293567, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.027000] env[62814]: DEBUG nova.compute.manager [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 768.027000] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.028054] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760e37cb-5a81-495d-9954-aafe352a3eaf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.037852] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 768.038789] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1056a913-9eb1-4f38-93d0-8735c35b6e53 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.047474] env[62814]: DEBUG oslo_vmware.api [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 768.047474] env[62814]: value = "task-4293568" [ 768.047474] env[62814]: _type = "Task" [ 768.047474] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.057896] env[62814]: DEBUG oslo_vmware.api [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293568, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.150133] env[62814]: DEBUG nova.compute.manager [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 768.150133] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.151193] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-738d83cc-a8db-44ae-a03a-157dc6f94a7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.162673] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 768.162978] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5a8342f8-cca5-43f7-9f73-21a630ed326e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.174215] env[62814]: DEBUG oslo_vmware.api [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 768.174215] env[62814]: value = "task-4293569" [ 768.174215] env[62814]: _type = "Task" [ 768.174215] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.186239] env[62814]: DEBUG oslo_vmware.api [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293569, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.273914] env[62814]: INFO nova.compute.manager [-] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Took 1.40 seconds to deallocate network for instance. [ 768.305926] env[62814]: DEBUG nova.scheduler.client.report [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 768.356060] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d7d3f97-cc17-4618-901c-6db0bd7b266b tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "1a16bd56-0992-4bec-bd3f-2836dfe68579" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.534s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.392590] env[62814]: DEBUG oslo_concurrency.lockutils [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.424484] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293567, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.791097} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.424778] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 9677b03f-7138-47b7-b1e8-f3714d11e550/9677b03f-7138-47b7-b1e8-f3714d11e550.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 768.425011] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 768.425292] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-34e5b52d-4286-4c3e-aac8-a2aecf533ca7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.433708] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 768.433708] env[62814]: value = "task-4293570" [ 768.433708] env[62814]: _type = "Task" [ 768.433708] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.443451] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293570, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.481121] env[62814]: DEBUG nova.compute.manager [req-6b032865-9be4-4fdc-b4ce-7acdf0180d8f req-1f62eca8-0bc5-4af8-a440-f178a8dbcbf3 service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received event network-vif-deleted-fbfe5ed2-1472-41c4-a028-38e81d37b3fe {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 768.481813] env[62814]: DEBUG nova.compute.manager [req-6b032865-9be4-4fdc-b4ce-7acdf0180d8f req-1f62eca8-0bc5-4af8-a440-f178a8dbcbf3 service nova] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Received event network-vif-deleted-ce7e8021-3ebf-4e8c-a0c2-618fa44c6957 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 768.481813] env[62814]: DEBUG nova.compute.manager [req-6b032865-9be4-4fdc-b4ce-7acdf0180d8f req-1f62eca8-0bc5-4af8-a440-f178a8dbcbf3 service nova] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Received event network-vif-deleted-2c623055-7773-4e8b-86c0-8765158b3620 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 768.518184] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquiring lock "1a16bd56-0992-4bec-bd3f-2836dfe68579" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.518184] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "1a16bd56-0992-4bec-bd3f-2836dfe68579" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.518184] env[62814]: INFO nova.compute.manager [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Rebooting instance [ 768.560439] env[62814]: DEBUG oslo_vmware.api [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293568, 'name': PowerOffVM_Task, 'duration_secs': 0.33537} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.561029] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 768.561029] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 768.561992] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2e836dc9-629e-4de8-a51d-388c1f6e8fd8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.634192] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 768.634557] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 768.634854] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Deleting the datastore file [datastore2] 961f313b-b43f-4531-8a4b-0a39421d6a34 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 768.637142] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-45df546e-b195-497b-9bd2-9af8413c24fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.644831] env[62814]: DEBUG oslo_vmware.api [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 768.644831] env[62814]: value = "task-4293572" [ 768.644831] env[62814]: _type = "Task" [ 768.644831] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.655228] env[62814]: DEBUG oslo_vmware.api [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293572, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.683619] env[62814]: DEBUG oslo_vmware.api [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293569, 'name': PowerOffVM_Task, 'duration_secs': 0.234349} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.683904] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 768.684085] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 768.684338] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e7c8060-7a2c-4c4f-87b9-f51fb47c1b4a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.778083] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 768.778326] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 768.778485] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Deleting the datastore file [datastore2] 64a475e4-6713-408b-a63a-a43b5fed5ec8 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 768.778753] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-576f34d1-0aa4-4d4c-a106-4e61b9324d2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.785621] env[62814]: DEBUG oslo_vmware.api [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for the task: (returnval){ [ 768.785621] env[62814]: value = "task-4293574" [ 768.785621] env[62814]: _type = "Task" [ 768.785621] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.789615] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.798127] env[62814]: DEBUG oslo_vmware.api [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293574, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.813200] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.779s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 768.813836] env[62814]: DEBUG nova.compute.manager [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 768.817268] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 34.342s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 768.859570] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 768.903123] env[62814]: DEBUG nova.objects.instance [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'flavor' on Instance uuid 59d21ef7-df97-49ac-9329-4c18df6dd087 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 768.949667] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293570, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070048} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.949667] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 768.950497] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dca7fe1-9d0d-47c5-9887-7e79baef6f03 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.979154] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] 9677b03f-7138-47b7-b1e8-f3714d11e550/9677b03f-7138-47b7-b1e8-f3714d11e550.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 768.986140] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f288eab-9655-4a3a-b826-b4ed65c86655 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.004160] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 769.004160] env[62814]: value = "task-4293575" [ 769.004160] env[62814]: _type = "Task" [ 769.004160] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.013839] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293575, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.054742] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquiring lock "refresh_cache-1a16bd56-0992-4bec-bd3f-2836dfe68579" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.054981] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquired lock "refresh_cache-1a16bd56-0992-4bec-bd3f-2836dfe68579" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.055710] env[62814]: DEBUG nova.network.neutron [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.157719] env[62814]: DEBUG oslo_vmware.api [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293572, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.332983} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.157993] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 769.158229] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 769.158373] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.158551] env[62814]: INFO nova.compute.manager [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Took 1.13 seconds to destroy the instance on the hypervisor. [ 769.158794] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 769.159085] env[62814]: DEBUG nova.compute.manager [-] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 769.159128] env[62814]: DEBUG nova.network.neutron [-] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.299336] env[62814]: DEBUG oslo_vmware.api [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Task: {'id': task-4293574, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.298602} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.299579] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 769.300133] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 769.300531] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.300820] env[62814]: INFO nova.compute.manager [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Took 1.15 seconds to destroy the instance on the hypervisor. [ 769.301234] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 769.302751] env[62814]: DEBUG nova.compute.manager [-] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 769.302751] env[62814]: DEBUG nova.network.neutron [-] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 769.321466] env[62814]: DEBUG nova.compute.utils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 769.336105] env[62814]: DEBUG nova.compute.manager [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Not allocating networking since 'none' was specified. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 769.387938] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.409125] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.409261] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 769.409451] env[62814]: DEBUG nova.network.neutron [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.409685] env[62814]: DEBUG nova.objects.instance [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'info_cache' on Instance uuid 59d21ef7-df97-49ac-9329-4c18df6dd087 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 769.517105] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293575, 'name': ReconfigVM_Task, 'duration_secs': 0.386388} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.517611] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Reconfigured VM instance instance-00000026 to attach disk [datastore2] 9677b03f-7138-47b7-b1e8-f3714d11e550/9677b03f-7138-47b7-b1e8-f3714d11e550.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 769.518416] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07dc7dbd-4891-4d0f-a4ca-82f4d30bd7fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.525210] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 769.525210] env[62814]: value = "task-4293576" [ 769.525210] env[62814]: _type = "Task" [ 769.525210] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.539312] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293576, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.686400] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquiring lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 769.686772] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.837149] env[62814]: DEBUG nova.compute.manager [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 769.888744] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 17bfe703-ff96-4cb7-8535-eed57e10d673 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.889060] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 5087b202-9bba-4489-823b-5d93cbf116e2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.889170] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance c7cef7f3-11db-44e1-a454-98830b465b52 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.889340] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 3d3f32fc-276a-49be-b471-01a5d6fc5069 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.889524] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d9242042-6209-4b04-bf00-00dd04d9d6a0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.889719] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 6a592192-1b41-4be2-84a6-c3b76a4e5643 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.889868] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 59d21ef7-df97-49ac-9329-4c18df6dd087 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.890207] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.890377] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 4ed66b36-b6c6-4673-9c03-169a01134574 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.890537] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance f5ad9d70-75fb-4881-8853-5ede4d0903f2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.890696] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 64a475e4-6713-408b-a63a-a43b5fed5ec8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.890850] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 961f313b-b43f-4531-8a4b-0a39421d6a34 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.891046] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance c7829a8b-a9f7-40b1-958c-732e2eaa2b3c is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 769.891273] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance bd0933ca-aab7-4dd4-a570-1a58a720f377 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.891445] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 369cd937-4c18-4068-ae59-70a1d585094b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 769.891744] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance a589a3d8-20dc-4ff5-a192-c540e29f39d6 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 769.891924] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 217e5812-95cc-4104-8d7b-82dfca2c0fcc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.892143] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 961dedc6-5ddd-4620-bfa7-54cf0c520af3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 769.892299] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 9343301c-dfe9-41b0-b4a0-067af544d297 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.892473] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 56f9e03c-1157-442f-8add-156627a6ec1e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 769.892844] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 1a16bd56-0992-4bec-bd3f-2836dfe68579 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.893754] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 9677b03f-7138-47b7-b1e8-f3714d11e550 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.893754] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance c9e11016-e92a-459e-b5ee-b0e43ce29450 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 769.917418] env[62814]: DEBUG nova.objects.base [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Object Instance<59d21ef7-df97-49ac-9329-4c18df6dd087> lazy-loaded attributes: flavor,info_cache {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 769.942848] env[62814]: DEBUG nova.network.neutron [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Updating instance_info_cache with network_info: [{"id": "02f411b5-731e-45d9-a7dc-9aca642cd9aa", "address": "fa:16:3e:55:12:ae", "network": {"id": "3cbd3440-c630-4c57-a5a9-d109b77a6064", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1749196117-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0e8c68a04eda4221925a5397578c0280", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c8ee8640-3787-4c27-9581-962ddb2be7e5", "external-id": "nsx-vlan-transportzone-224", "segmentation_id": 224, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap02f411b5-73", "ovs_interfaceid": "02f411b5-731e-45d9-a7dc-9aca642cd9aa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.036192] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293576, 'name': Rename_Task, 'duration_secs': 0.162594} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.036482] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 770.036725] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1e590ef1-9c48-41d6-9c00-87061a533978 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.045222] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 770.045222] env[62814]: value = "task-4293577" [ 770.045222] env[62814]: _type = "Task" [ 770.045222] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.055679] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.268806] env[62814]: DEBUG nova.network.neutron [-] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.386278] env[62814]: DEBUG nova.network.neutron [-] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.397561] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance b289800f-49ce-49e7-b6bc-a3b4ec84b434 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.446172] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Releasing lock "refresh_cache-1a16bd56-0992-4bec-bd3f-2836dfe68579" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 770.517965] env[62814]: DEBUG nova.compute.manager [req-dddfdca6-210f-46b3-9f48-a4a9b2dc8178 req-842fbfb3-59fa-4b56-bcde-f4fe462cb20f service nova] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Received event network-vif-deleted-cb1f398a-cfcd-4f37-8fb0-75ada6aafe7b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 770.518838] env[62814]: DEBUG nova.compute.manager [req-dddfdca6-210f-46b3-9f48-a4a9b2dc8178 req-842fbfb3-59fa-4b56-bcde-f4fe462cb20f service nova] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Received event network-vif-deleted-33a51a23-d1ad-456b-9a36-3c0a6df8f966 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 770.557324] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293577, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.744645] env[62814]: DEBUG nova.network.neutron [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance_info_cache with network_info: [{"id": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "address": "fa:16:3e:93:37:3e", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa00402c1-eb", "ovs_interfaceid": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.771280] env[62814]: INFO nova.compute.manager [-] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Took 1.61 seconds to deallocate network for instance. [ 770.854817] env[62814]: DEBUG nova.compute.manager [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 770.886556] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 770.886891] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.887395] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 770.887680] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.887877] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 770.888093] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 770.888441] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 770.888680] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 770.888927] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 770.889235] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 770.889634] env[62814]: DEBUG nova.virt.hardware [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 770.890384] env[62814]: INFO nova.compute.manager [-] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Took 1.59 seconds to deallocate network for instance. [ 770.891379] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1868d17c-03bb-4acf-8cbe-825251223442 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.902229] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d366a755-49b4-427b-8564-d8572a7fbbb7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.906534] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c77803c-3f08-44df-ad44-5b423be09a01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.922024] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 770.927520] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Creating folder: Project (3a6053a4ce7e4779a110fa0fa9f9b0ae). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 770.928514] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-483f5d01-7ca4-4832-be86-d046631456f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.939486] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Created folder: Project (3a6053a4ce7e4779a110fa0fa9f9b0ae) in parent group-v845547. [ 770.939676] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Creating folder: Instances. Parent ref: group-v845656. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 770.939920] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4bc26c9a-0bff-4dae-84de-510a6e409712 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.953159] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Created folder: Instances in parent group-v845656. [ 770.954626] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 770.954626] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 770.954626] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-44826d27-6566-4895-837f-66f5883c4e3c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.968635] env[62814]: DEBUG nova.compute.manager [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 770.970349] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41b6479a-7ab4-4ee5-be25-ab43ccd05dca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.987024] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 770.987024] env[62814]: value = "task-4293580" [ 770.987024] env[62814]: _type = "Task" [ 770.987024] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.997535] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293580, 'name': CreateVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.058883] env[62814]: DEBUG oslo_vmware.api [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293577, 'name': PowerOnVM_Task, 'duration_secs': 0.564589} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.059349] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 771.059675] env[62814]: INFO nova.compute.manager [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Took 9.69 seconds to spawn the instance on the hypervisor. [ 771.059959] env[62814]: DEBUG nova.compute.manager [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 771.061240] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3210aa33-3f44-4da5-a6cc-573b1252b390 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.247930] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.278312] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.413763] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.415757] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 127fee64-fd56-4a23-bdd2-18c817898fd5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.507804] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293580, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.586321] env[62814]: INFO nova.compute.manager [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Took 41.95 seconds to build instance. [ 771.919966] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 3e4479a0-8edd-4b37-8cc9-2c91275b88ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.002812] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be676e9-51a0-419b-b120-4fe0afc0ded0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.010962] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293580, 'name': CreateVM_Task, 'duration_secs': 0.630975} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.011282] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 772.011709] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.011843] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.012234] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 772.012436] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10270dca-3a41-4c7d-adeb-1ec840b5cd2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.018161] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Doing hard reboot of VM {{(pid=62814) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 772.018161] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-3e6e4b67-44b7-4c4d-bbcf-97e776f58080 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.019511] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 772.019511] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e55039-138d-59d2-ea23-c61c01d549b9" [ 772.019511] env[62814]: _type = "Task" [ 772.019511] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.024119] env[62814]: DEBUG oslo_vmware.api [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 772.024119] env[62814]: value = "task-4293581" [ 772.024119] env[62814]: _type = "Task" [ 772.024119] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.030855] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e55039-138d-59d2-ea23-c61c01d549b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.037991] env[62814]: DEBUG oslo_vmware.api [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293581, 'name': ResetVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.092342] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72b22c9d-e261-45e2-9cf9-6ec8bf83aff9 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lock "9677b03f-7138-47b7-b1e8-f3714d11e550" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.457s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 772.254640] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 772.254750] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45ecf07d-c5f6-4831-bef0-23b8de06ff87 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.262655] env[62814]: DEBUG oslo_vmware.api [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 772.262655] env[62814]: value = "task-4293582" [ 772.262655] env[62814]: _type = "Task" [ 772.262655] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.271865] env[62814]: DEBUG oslo_vmware.api [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293582, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.424623] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance cbc5985a-38e7-4e52-9fb0-264b5cec013a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.534614] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e55039-138d-59d2-ea23-c61c01d549b9, 'name': SearchDatastore_Task, 'duration_secs': 0.017355} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.535961] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.536240] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.536479] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.536658] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 772.536828] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.537355] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7855b66-50f7-4086-8fe6-864024655b54 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.543247] env[62814]: DEBUG oslo_vmware.api [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293581, 'name': ResetVM_Task, 'duration_secs': 0.196891} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.545034] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Did hard reboot of VM {{(pid=62814) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 772.545034] env[62814]: DEBUG nova.compute.manager [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 772.545034] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14e0e74-dca6-4cad-bf98-0a3504bfb83a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.551146] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.551146] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 772.552745] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d77933ed-39da-4ff5-b3c6-51368f51a8ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.562819] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 772.562819] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52121cb5-3761-0d18-744b-f1469d076bc7" [ 772.562819] env[62814]: _type = "Task" [ 772.562819] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.572158] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52121cb5-3761-0d18-744b-f1469d076bc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.595031] env[62814]: DEBUG nova.compute.manager [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 772.775397] env[62814]: DEBUG oslo_vmware.api [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293582, 'name': PowerOnVM_Task, 'duration_secs': 0.393525} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.775795] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 772.776171] env[62814]: DEBUG nova.compute.manager [None req-c04307b8-6edf-4dea-bb31-2fb4a60ac45f tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 772.776927] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e90b79-ab88-4d4a-b25f-5d1cc9729cf2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.787092] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "425e8edd-c002-45a1-bb6f-ee3ac8812509" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.787092] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "425e8edd-c002-45a1-bb6f-ee3ac8812509" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.927632] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 1f40dc62-1a58-4cfb-8785-c37b68747f37 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.064714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc534057-b7f5-4c18-8f38-94eab00882e5 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "1a16bd56-0992-4bec-bd3f-2836dfe68579" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.548s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.077572] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52121cb5-3761-0d18-744b-f1469d076bc7, 'name': SearchDatastore_Task, 'duration_secs': 0.03761} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.078666] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-733bb979-83e5-46cb-962a-edf62ae46d47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.085387] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 773.085387] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f8728b-0ea4-32f2-7d5d-b9db5706d044" [ 773.085387] env[62814]: _type = "Task" [ 773.085387] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.094533] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f8728b-0ea4-32f2-7d5d-b9db5706d044, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.127137] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.434561] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 2d0b24c5-3593-4ef2-a637-d3590242ad79 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.530154] env[62814]: DEBUG nova.compute.manager [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 773.531106] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe25fe49-7a73-44f4-b6da-a01f9ebf09e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.595938] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f8728b-0ea4-32f2-7d5d-b9db5706d044, 'name': SearchDatastore_Task, 'duration_secs': 0.01638} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.596237] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.596539] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c9e11016-e92a-459e-b5ee-b0e43ce29450/c9e11016-e92a-459e-b5ee-b0e43ce29450.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.596742] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df934c55-695d-4cb6-99d0-1d4686599cd1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.603456] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 773.603456] env[62814]: value = "task-4293583" [ 773.603456] env[62814]: _type = "Task" [ 773.603456] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.612878] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293583, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.686156] env[62814]: DEBUG oslo_vmware.rw_handles [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a257d-7a25-e314-6f4b-e79eb6c30a9c/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 773.687096] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104f8bc5-4cc2-476c-bc69-9b83753203f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.693331] env[62814]: DEBUG oslo_vmware.rw_handles [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a257d-7a25-e314-6f4b-e79eb6c30a9c/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 773.693486] env[62814]: ERROR oslo_vmware.rw_handles [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a257d-7a25-e314-6f4b-e79eb6c30a9c/disk-0.vmdk due to incomplete transfer. [ 773.693731] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-eb6930c5-4066-4b1e-9c1b-7de2d37f3cc9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.701014] env[62814]: DEBUG oslo_vmware.rw_handles [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a257d-7a25-e314-6f4b-e79eb6c30a9c/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 773.701219] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Uploaded image f5e1a700-0673-4eca-adf9-4520693e9282 to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 773.703302] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 773.703539] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e662553e-e460-4201-902a-1f348662ffbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.709315] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 773.709315] env[62814]: value = "task-4293584" [ 773.709315] env[62814]: _type = "Task" [ 773.709315] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.716996] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293584, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.741747] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquiring lock "1a16bd56-0992-4bec-bd3f-2836dfe68579" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.742113] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "1a16bd56-0992-4bec-bd3f-2836dfe68579" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.742419] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquiring lock "1a16bd56-0992-4bec-bd3f-2836dfe68579-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.742755] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "1a16bd56-0992-4bec-bd3f-2836dfe68579-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 773.743045] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "1a16bd56-0992-4bec-bd3f-2836dfe68579-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 773.746503] env[62814]: INFO nova.compute.manager [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Terminating instance [ 773.938833] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance c18acec7-cf95-4cdf-aa49-32419d364534 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.043042] env[62814]: INFO nova.compute.manager [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] instance snapshotting [ 774.047185] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3993304a-f494-4edc-8c65-81022cde6196 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.071372] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2694b5-718f-43ad-b5f3-0a49970908ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.114665] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293583, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.221586] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293584, 'name': Destroy_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.251644] env[62814]: DEBUG nova.compute.manager [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 774.252011] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 774.253392] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782596c3-7ed9-42d8-b7cc-a41993420a32 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.261839] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 774.261936] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cd7c8e0a-c7a1-4393-898b-2557cb2aa88f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.268211] env[62814]: DEBUG oslo_vmware.api [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 774.268211] env[62814]: value = "task-4293585" [ 774.268211] env[62814]: _type = "Task" [ 774.268211] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.277661] env[62814]: DEBUG oslo_vmware.api [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293585, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.442075] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 9cd89da4-d190-4bfc-81e1-da15b98d54ba has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.586714] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 774.587111] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-040b9442-c1d6-4c72-8b65-5a56abf3aa47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.595418] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 774.595418] env[62814]: value = "task-4293586" [ 774.595418] env[62814]: _type = "Task" [ 774.595418] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.603080] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293586, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.614638] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293583, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.637871} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.615012] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c9e11016-e92a-459e-b5ee-b0e43ce29450/c9e11016-e92a-459e-b5ee-b0e43ce29450.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 774.615295] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.615638] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c68743c4-36b2-42c0-9837-c374d53a6910 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.621448] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 774.621448] env[62814]: value = "task-4293587" [ 774.621448] env[62814]: _type = "Task" [ 774.621448] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.630414] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.719512] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293584, 'name': Destroy_Task, 'duration_secs': 0.630461} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.719809] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Destroyed the VM [ 774.720035] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 774.720296] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-954cb76e-1262-4f51-b7a9-eefc08ba4aff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.726666] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 774.726666] env[62814]: value = "task-4293588" [ 774.726666] env[62814]: _type = "Task" [ 774.726666] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.736281] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293588, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.777521] env[62814]: DEBUG oslo_vmware.api [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293585, 'name': PowerOffVM_Task, 'duration_secs': 0.183698} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.777632] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 774.777745] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 774.777988] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b2bf6f74-5220-45c8-aaef-4546749a6950 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.833680] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 774.834539] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 774.834539] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Deleting the datastore file [datastore2] 1a16bd56-0992-4bec-bd3f-2836dfe68579 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 774.834539] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-033c546d-ec5d-4935-8e9c-947a218f58b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.842044] env[62814]: DEBUG oslo_vmware.api [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for the task: (returnval){ [ 774.842044] env[62814]: value = "task-4293590" [ 774.842044] env[62814]: _type = "Task" [ 774.842044] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.851754] env[62814]: DEBUG oslo_vmware.api [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293590, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.945433] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 7cca1c35-6bfc-450d-ba74-0e825b160e8f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.105305] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293586, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.130705] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063932} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.130928] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.131724] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5def79a6-27f8-488f-83b4-b58d20548d5f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.151206] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] c9e11016-e92a-459e-b5ee-b0e43ce29450/c9e11016-e92a-459e-b5ee-b0e43ce29450.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.151483] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7da1e0c-9b20-4139-8c01-71978de756fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.174325] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 775.174325] env[62814]: value = "task-4293591" [ 775.174325] env[62814]: _type = "Task" [ 775.174325] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.183162] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293591, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.236711] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293588, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.328292] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.328292] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 775.328292] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.328292] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 775.328292] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.328551] env[62814]: INFO nova.compute.manager [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Terminating instance [ 775.355457] env[62814]: DEBUG oslo_vmware.api [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Task: {'id': task-4293590, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156659} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.355798] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 775.356274] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 775.356274] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 775.356457] env[62814]: INFO nova.compute.manager [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Took 1.10 seconds to destroy the instance on the hypervisor. [ 775.356750] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 775.356897] env[62814]: DEBUG nova.compute.manager [-] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 775.356959] env[62814]: DEBUG nova.network.neutron [-] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 775.452355] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 413eaa2e-7bbc-402e-b0d3-f030b6395d7b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.452640] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 18 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 775.452963] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4032MB phys_disk=149GB used_disk=17GB total_vcpus=48 used_vcpus=18 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 775.605448] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293586, 'name': CreateSnapshot_Task, 'duration_secs': 0.906186} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.607962] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 775.608861] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c805d485-e928-442f-a4e3-0c3b9214c679 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.676557] env[62814]: DEBUG nova.compute.manager [req-8f896a8c-7dd3-49b3-a041-d936308010dd req-1522ab0c-b533-4c7b-a784-15058b26c200 service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Received event network-vif-deleted-02f411b5-731e-45d9-a7dc-9aca642cd9aa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 775.676557] env[62814]: INFO nova.compute.manager [req-8f896a8c-7dd3-49b3-a041-d936308010dd req-1522ab0c-b533-4c7b-a784-15058b26c200 service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Neutron deleted interface 02f411b5-731e-45d9-a7dc-9aca642cd9aa; detaching it from the instance and deleting it from the info cache [ 775.676557] env[62814]: DEBUG nova.network.neutron [req-8f896a8c-7dd3-49b3-a041-d936308010dd req-1522ab0c-b533-4c7b-a784-15058b26c200 service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.691857] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293591, 'name': ReconfigVM_Task, 'duration_secs': 0.38102} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.691976] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Reconfigured VM instance instance-00000027 to attach disk [datastore2] c9e11016-e92a-459e-b5ee-b0e43ce29450/c9e11016-e92a-459e-b5ee-b0e43ce29450.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 775.692627] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13f2d584-992e-4595-83e4-925d4b576899 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.700627] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 775.700627] env[62814]: value = "task-4293592" [ 775.700627] env[62814]: _type = "Task" [ 775.700627] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.711717] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293592, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.739226] env[62814]: DEBUG oslo_vmware.api [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293588, 'name': RemoveSnapshot_Task, 'duration_secs': 0.710448} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.739619] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 775.739878] env[62814]: INFO nova.compute.manager [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Took 17.62 seconds to snapshot the instance on the hypervisor. [ 775.827296] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "c8765756-0870-4a06-a1a5-d02177959b29" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.827538] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "c8765756-0870-4a06-a1a5-d02177959b29" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 775.836094] env[62814]: DEBUG nova.compute.manager [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 775.836317] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 775.837248] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904574e3-0cf2-4992-a0ee-9c58b4d46f71 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.848968] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 775.849253] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80592d2a-772f-4693-b820-23744ba525be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.907193] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 775.907529] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 775.908198] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleting the datastore file [datastore2] 217e5812-95cc-4104-8d7b-82dfca2c0fcc {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 775.908449] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-699aa809-ae63-433a-8061-f9a3e08e2e28 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.922071] env[62814]: DEBUG oslo_vmware.api [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 775.922071] env[62814]: value = "task-4293594" [ 775.922071] env[62814]: _type = "Task" [ 775.922071] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.932757] env[62814]: DEBUG oslo_vmware.api [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293594, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.003793] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba4ecfe-4827-44dc-b841-801bd740121a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.011441] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-390352b5-8802-4712-bf25-1c009800031b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.042885] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e451fe5-a7f0-40bb-a303-5bc4dfdc0968 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.050293] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d4b4d2-8c03-4934-b9f3-39e0b6c55b83 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.065604] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.130199] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 776.130569] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2cb83d6c-c299-4dc8-b7d3-c026dade1bcd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.139929] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 776.139929] env[62814]: value = "task-4293595" [ 776.139929] env[62814]: _type = "Task" [ 776.139929] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.147541] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293595, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.148897] env[62814]: DEBUG nova.network.neutron [-] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.179967] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-62d2ad76-ad54-4bb9-8e8c-e738bd5aaae1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.189534] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4187e91c-209d-48e4-92c3-74cd3904ef64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.208899] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293592, 'name': Rename_Task, 'duration_secs': 0.151777} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.220482] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 776.220889] env[62814]: DEBUG nova.compute.manager [req-8f896a8c-7dd3-49b3-a041-d936308010dd req-1522ab0c-b533-4c7b-a784-15058b26c200 service nova] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Detach interface failed, port_id=02f411b5-731e-45d9-a7dc-9aca642cd9aa, reason: Instance 1a16bd56-0992-4bec-bd3f-2836dfe68579 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 776.221832] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe5a32e0-7e31-4bf3-9347-86e59faa1660 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.227217] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 776.227217] env[62814]: value = "task-4293596" [ 776.227217] env[62814]: _type = "Task" [ 776.227217] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.234962] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293596, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.246656] env[62814]: DEBUG nova.compute.manager [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Instance disappeared during snapshot {{(pid=62814) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4595}} [ 776.260225] env[62814]: DEBUG nova.compute.manager [None req-752b41a8-7abd-4b67-90f1-3f766e80a3ae tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image not found during clean up f5e1a700-0673-4eca-adf9-4520693e9282 {{(pid=62814) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4601}} [ 776.434653] env[62814]: DEBUG oslo_vmware.api [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293594, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158782} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.434958] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 776.435199] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 776.435429] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 776.435654] env[62814]: INFO nova.compute.manager [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Took 0.60 seconds to destroy the instance on the hypervisor. [ 776.435944] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 776.436351] env[62814]: DEBUG nova.compute.manager [-] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 776.436464] env[62814]: DEBUG nova.network.neutron [-] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 776.568940] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 776.649518] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293595, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.651062] env[62814]: INFO nova.compute.manager [-] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Took 1.29 seconds to deallocate network for instance. [ 776.737018] env[62814]: DEBUG oslo_vmware.api [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293596, 'name': PowerOnVM_Task, 'duration_secs': 0.438227} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.737213] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 776.737416] env[62814]: INFO nova.compute.manager [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Took 5.88 seconds to spawn the instance on the hypervisor. [ 776.737591] env[62814]: DEBUG nova.compute.manager [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 776.738365] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6090fe-5929-4999-ba93-2bb29f087f5e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.074689] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 777.074982] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.258s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.075271] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 40.380s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.075454] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.078137] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.861s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.079656] env[62814]: INFO nova.compute.claims [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.106907] env[62814]: INFO nova.scheduler.client.report [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Deleted allocations for instance c7829a8b-a9f7-40b1-958c-732e2eaa2b3c [ 777.151422] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293595, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.157328] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 777.184759] env[62814]: DEBUG nova.network.neutron [-] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.256460] env[62814]: INFO nova.compute.manager [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Took 42.97 seconds to build instance. [ 777.614672] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2285d1a1-0893-43a0-8ed8-a5d424e0b560 tempest-InstanceActionsV221TestJSON-725096177 tempest-InstanceActionsV221TestJSON-725096177-project-member] Lock "c7829a8b-a9f7-40b1-958c-732e2eaa2b3c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 44.357s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.652077] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293595, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.687338] env[62814]: INFO nova.compute.manager [-] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Took 1.25 seconds to deallocate network for instance. [ 777.698757] env[62814]: DEBUG nova.compute.manager [req-1b5cc2e0-77f7-49d1-ad3e-8143fe223153 req-c590b16e-af82-4927-97ac-ed8e048ad9ef service nova] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Received event network-vif-deleted-51cf8a1d-3070-4968-bd29-37c3fa579184 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 777.758274] env[62814]: DEBUG oslo_concurrency.lockutils [None req-877a7f3e-0fb0-4922-859a-bbde2d46c55b tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "c9e11016-e92a-459e-b5ee-b0e43ce29450" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.360s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 778.095218] env[62814]: INFO nova.compute.manager [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Rebuilding instance [ 778.136994] env[62814]: DEBUG nova.compute.manager [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 778.138026] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8e9db1-0c33-4d0c-ab47-a5f8fbfd8a18 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.156271] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293595, 'name': CloneVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.195394] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.260511] env[62814]: DEBUG nova.compute.manager [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 778.570588] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b49d331-cef3-4092-9a19-7f31057762f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.578557] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b2b4dd-791e-47f0-99b3-714b22a51560 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.609280] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777f12f3-748e-4d98-b7cc-4d295d23e5a3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.618149] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d9da40-176b-4ab3-aad9-ec257bdfde5c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.635023] env[62814]: DEBUG nova.compute.provider_tree [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 778.658731] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293595, 'name': CloneVM_Task, 'duration_secs': 2.085095} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.660105] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Created linked-clone VM from snapshot [ 778.661198] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99373b5-ff5f-44ec-90f9-08f9af008899 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.670034] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Uploading image c73af40e-dd5c-4c62-8e27-a5a56963e1c7 {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 778.700233] env[62814]: DEBUG oslo_vmware.rw_handles [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 778.700233] env[62814]: value = "vm-845660" [ 778.700233] env[62814]: _type = "VirtualMachine" [ 778.700233] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 778.700233] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-eaf12044-0a60-4d03-bda6-86bd742f9abd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.706133] env[62814]: DEBUG oslo_vmware.rw_handles [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lease: (returnval){ [ 778.706133] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5251e28d-8236-7391-e59c-6e0645031b7f" [ 778.706133] env[62814]: _type = "HttpNfcLease" [ 778.706133] env[62814]: } obtained for exporting VM: (result){ [ 778.706133] env[62814]: value = "vm-845660" [ 778.706133] env[62814]: _type = "VirtualMachine" [ 778.706133] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 778.706433] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the lease: (returnval){ [ 778.706433] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5251e28d-8236-7391-e59c-6e0645031b7f" [ 778.706433] env[62814]: _type = "HttpNfcLease" [ 778.706433] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 778.713834] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 778.713834] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5251e28d-8236-7391-e59c-6e0645031b7f" [ 778.713834] env[62814]: _type = "HttpNfcLease" [ 778.713834] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 778.790633] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.157981] env[62814]: ERROR nova.scheduler.client.report [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [req-fc5a08d1-be2c-437c-a2cc-36a6fadfdfa4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-fc5a08d1-be2c-437c-a2cc-36a6fadfdfa4"}]} [ 779.165091] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 779.170324] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7059b4e-fb48-4501-b3e6-3970102f0e15 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.178720] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 779.178720] env[62814]: value = "task-4293598" [ 779.178720] env[62814]: _type = "Task" [ 779.178720] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.185162] env[62814]: DEBUG nova.scheduler.client.report [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 779.195452] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293598, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.201545] env[62814]: DEBUG nova.scheduler.client.report [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 779.201820] env[62814]: DEBUG nova.compute.provider_tree [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 779.218036] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 779.218036] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5251e28d-8236-7391-e59c-6e0645031b7f" [ 779.218036] env[62814]: _type = "HttpNfcLease" [ 779.218036] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 779.218965] env[62814]: DEBUG nova.scheduler.client.report [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 779.220934] env[62814]: DEBUG oslo_vmware.rw_handles [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 779.220934] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5251e28d-8236-7391-e59c-6e0645031b7f" [ 779.220934] env[62814]: _type = "HttpNfcLease" [ 779.220934] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 779.221856] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1344fbb-871b-4127-8d38-92e692d3f4e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.233929] env[62814]: DEBUG oslo_vmware.rw_handles [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ad171e-cac5-2a80-441e-698d3f02d9df/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 779.234142] env[62814]: DEBUG oslo_vmware.rw_handles [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ad171e-cac5-2a80-441e-698d3f02d9df/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 779.292578] env[62814]: DEBUG nova.scheduler.client.report [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 779.325228] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2963fc5d-eb4d-4cbc-a991-2d40b04f640a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.691939] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293598, 'name': PowerOffVM_Task, 'duration_secs': 0.184213} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.692191] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 779.692422] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 779.693302] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0608c5a4-3d35-46be-84b1-75cb20de1a27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.706063] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 779.706063] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-387cec48-01c8-46d2-bbfc-5aedad84538d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.734743] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 779.735139] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 779.736490] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Deleting the datastore file [datastore2] c9e11016-e92a-459e-b5ee-b0e43ce29450 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 779.736993] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2458844d-703e-4d8e-b2a6-c1f0f90da136 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.751830] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 779.751830] env[62814]: value = "task-4293600" [ 779.751830] env[62814]: _type = "Task" [ 779.751830] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.767557] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293600, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.923316] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4223e57f-5433-4207-8bf6-608e935e0eac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.930932] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c0d76f9-07c0-4ee7-ba5a-28bfcbf001a2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.966837] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faaccdb9-7964-447d-881d-fad4b828beb8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.975608] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec6c27f-d6b8-4783-883c-4225aceca2ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.992467] env[62814]: DEBUG nova.compute.provider_tree [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 780.268936] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293600, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13801} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.269371] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 780.269503] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 780.269768] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 780.541073] env[62814]: DEBUG nova.scheduler.client.report [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 66 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 780.541073] env[62814]: DEBUG nova.compute.provider_tree [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 66 to 67 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 780.541073] env[62814]: DEBUG nova.compute.provider_tree [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 781.048840] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.969s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 781.048840] env[62814]: DEBUG nova.compute.manager [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 781.051419] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.718s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 781.056926] env[62814]: INFO nova.compute.claims [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.308391] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 781.308728] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.308774] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 781.308952] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.309115] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 781.309277] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 781.309591] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 781.309775] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 781.309962] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 781.310129] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 781.310306] env[62814]: DEBUG nova.virt.hardware [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 781.311279] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bf5bc41-0d2b-40a6-a231-246379883e0b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.319247] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66192712-84cf-4905-951b-050ed04b8423 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.333063] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.338627] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 781.338890] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 781.339116] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dddfa0ac-8e07-418f-8162-922f8ff795c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.356503] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.356503] env[62814]: value = "task-4293601" [ 781.356503] env[62814]: _type = "Task" [ 781.356503] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.364759] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293601, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.561919] env[62814]: DEBUG nova.compute.utils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 781.565758] env[62814]: DEBUG nova.compute.manager [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 781.566023] env[62814]: DEBUG nova.network.neutron [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 781.641883] env[62814]: DEBUG nova.policy [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ebebbb436074bd9b6d3b423d3740465', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '447bee1ea28948ebaa134d807e481edc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 781.869569] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293601, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.067218] env[62814]: DEBUG nova.compute.manager [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 782.097357] env[62814]: DEBUG nova.network.neutron [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Successfully created port: 19fc1ad9-b48d-45ba-9f90-85b3a5910021 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.368828] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293601, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.638498] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4a3d30-1500-4597-8a2b-89c19475785a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.647132] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88af7d7f-5455-4282-aa4a-eeac5c45f322 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.679142] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a2acec-96a3-4d1c-a407-a68f5b108de0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.686777] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd5a359f-928f-4d9e-a5d9-69d8d4b98fb6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.700618] env[62814]: DEBUG nova.compute.provider_tree [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.872487] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293601, 'name': CreateVM_Task, 'duration_secs': 1.344529} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.872487] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 782.872487] env[62814]: DEBUG oslo_concurrency.lockutils [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.872487] env[62814]: DEBUG oslo_concurrency.lockutils [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 782.872487] env[62814]: DEBUG oslo_concurrency.lockutils [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 782.872487] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c854c65a-df63-44ff-9e41-ff5eb9415e42 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.879285] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 782.879285] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f73c45-4d22-e884-3a6a-09065a843768" [ 782.879285] env[62814]: _type = "Task" [ 782.879285] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.890558] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f73c45-4d22-e884-3a6a-09065a843768, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.085550] env[62814]: DEBUG nova.compute.manager [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 783.123946] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 783.124275] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 783.124430] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 783.124613] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 783.124778] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 783.124936] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 783.125172] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 783.125342] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 783.125518] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 783.125682] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 783.125884] env[62814]: DEBUG nova.virt.hardware [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 783.126813] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-069dc62f-c7d9-42cb-97a5-0dd9587aa464 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.136115] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32af5e1e-f5cf-406a-a1ce-cfefae07859b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.207887] env[62814]: DEBUG nova.scheduler.client.report [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 783.317052] env[62814]: DEBUG oslo_concurrency.lockutils [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquiring lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.317425] env[62814]: DEBUG oslo_concurrency.lockutils [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.317739] env[62814]: DEBUG oslo_concurrency.lockutils [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquiring lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.317935] env[62814]: DEBUG oslo_concurrency.lockutils [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.318120] env[62814]: DEBUG oslo_concurrency.lockutils [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.320357] env[62814]: INFO nova.compute.manager [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Terminating instance [ 783.388385] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f73c45-4d22-e884-3a6a-09065a843768, 'name': SearchDatastore_Task, 'duration_secs': 0.012703} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.388890] env[62814]: DEBUG oslo_concurrency.lockutils [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.389084] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.389191] env[62814]: DEBUG oslo_concurrency.lockutils [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.389333] env[62814]: DEBUG oslo_concurrency.lockutils [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.389514] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.389854] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-364c6bbd-c5a1-45ad-aa04-6b057842e55c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.398488] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.398678] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.399418] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fd9fbce-9df8-4876-bc77-ddc099e82a75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.405029] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 783.405029] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527135dc-17e7-b09e-30bc-178a4803ff27" [ 783.405029] env[62814]: _type = "Task" [ 783.405029] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.412523] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527135dc-17e7-b09e-30bc-178a4803ff27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.713502] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.662s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.714209] env[62814]: DEBUG nova.compute.manager [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 783.717119] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.353s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.719357] env[62814]: INFO nova.compute.claims [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.724422] env[62814]: DEBUG nova.compute.manager [req-110eac04-6e20-442d-8cb6-9ccbf7d39b60 req-6d23286d-3b28-4ba8-afa4-ec2ffb262e68 service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Received event network-vif-plugged-19fc1ad9-b48d-45ba-9f90-85b3a5910021 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 783.724643] env[62814]: DEBUG oslo_concurrency.lockutils [req-110eac04-6e20-442d-8cb6-9ccbf7d39b60 req-6d23286d-3b28-4ba8-afa4-ec2ffb262e68 service nova] Acquiring lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.724838] env[62814]: DEBUG oslo_concurrency.lockutils [req-110eac04-6e20-442d-8cb6-9ccbf7d39b60 req-6d23286d-3b28-4ba8-afa4-ec2ffb262e68 service nova] Lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.725109] env[62814]: DEBUG oslo_concurrency.lockutils [req-110eac04-6e20-442d-8cb6-9ccbf7d39b60 req-6d23286d-3b28-4ba8-afa4-ec2ffb262e68 service nova] Lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.725241] env[62814]: DEBUG nova.compute.manager [req-110eac04-6e20-442d-8cb6-9ccbf7d39b60 req-6d23286d-3b28-4ba8-afa4-ec2ffb262e68 service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] No waiting events found dispatching network-vif-plugged-19fc1ad9-b48d-45ba-9f90-85b3a5910021 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 783.725417] env[62814]: WARNING nova.compute.manager [req-110eac04-6e20-442d-8cb6-9ccbf7d39b60 req-6d23286d-3b28-4ba8-afa4-ec2ffb262e68 service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Received unexpected event network-vif-plugged-19fc1ad9-b48d-45ba-9f90-85b3a5910021 for instance with vm_state building and task_state spawning. [ 783.802733] env[62814]: DEBUG nova.network.neutron [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Successfully updated port: 19fc1ad9-b48d-45ba-9f90-85b3a5910021 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 783.825087] env[62814]: DEBUG nova.compute.manager [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 783.825390] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 783.826252] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb1ebe6-255c-43cc-a800-9115b20d99bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.836430] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 783.837671] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ce37689-8ee9-4337-836c-4a489cac0c1d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.849031] env[62814]: DEBUG oslo_vmware.api [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for the task: (returnval){ [ 783.849031] env[62814]: value = "task-4293602" [ 783.849031] env[62814]: _type = "Task" [ 783.849031] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.861455] env[62814]: DEBUG oslo_vmware.api [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293602, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.917927] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527135dc-17e7-b09e-30bc-178a4803ff27, 'name': SearchDatastore_Task, 'duration_secs': 0.009971} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.918899] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-440ad18c-654c-4c52-ae1e-ed723430bd6f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.926240] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 783.926240] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52748a5a-14a9-f104-6fca-b906da5ed43a" [ 783.926240] env[62814]: _type = "Task" [ 783.926240] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.935846] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52748a5a-14a9-f104-6fca-b906da5ed43a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.225870] env[62814]: DEBUG nova.compute.utils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 784.229166] env[62814]: DEBUG nova.compute.manager [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 784.229528] env[62814]: DEBUG nova.network.neutron [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 784.273300] env[62814]: DEBUG nova.policy [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '356f22e207bb4c3d842d4c626473685d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fd6254d7e15b4c5ba267e8da03d1b82d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 784.306455] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquiring lock "refresh_cache-b289800f-49ce-49e7-b6bc-a3b4ec84b434" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.306687] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquired lock "refresh_cache-b289800f-49ce-49e7-b6bc-a3b4ec84b434" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 784.306824] env[62814]: DEBUG nova.network.neutron [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 784.363977] env[62814]: DEBUG oslo_vmware.api [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293602, 'name': PowerOffVM_Task, 'duration_secs': 0.218718} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.363977] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 784.364313] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 784.367549] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77431ce9-d58e-4818-ad28-2c7748032e75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.428236] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 784.428236] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 784.428569] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Deleting the datastore file [datastore2] 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 784.432419] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9cdf450a-04ce-4778-9035-51192368cae7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.439935] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52748a5a-14a9-f104-6fca-b906da5ed43a, 'name': SearchDatastore_Task, 'duration_secs': 0.014658} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.442678] env[62814]: DEBUG oslo_concurrency.lockutils [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 784.443073] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c9e11016-e92a-459e-b5ee-b0e43ce29450/c9e11016-e92a-459e-b5ee-b0e43ce29450.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 784.443617] env[62814]: DEBUG oslo_vmware.api [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for the task: (returnval){ [ 784.443617] env[62814]: value = "task-4293604" [ 784.443617] env[62814]: _type = "Task" [ 784.443617] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.443827] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3241f8e5-cad6-44e4-a9b7-04510b90d92e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.455531] env[62814]: DEBUG oslo_vmware.api [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293604, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.456958] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 784.456958] env[62814]: value = "task-4293605" [ 784.456958] env[62814]: _type = "Task" [ 784.456958] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.465706] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.664750] env[62814]: DEBUG nova.network.neutron [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Successfully created port: 1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 784.727187] env[62814]: DEBUG nova.compute.manager [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 784.869461] env[62814]: DEBUG nova.network.neutron [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.965665] env[62814]: DEBUG oslo_vmware.api [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Task: {'id': task-4293604, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176919} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.965952] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 784.966156] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 784.966336] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.966507] env[62814]: INFO nova.compute.manager [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Took 1.14 seconds to destroy the instance on the hypervisor. [ 784.966752] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 784.967768] env[62814]: DEBUG nova.compute.manager [-] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 784.967768] env[62814]: DEBUG nova.network.neutron [-] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.977897] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293605, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.149364] env[62814]: DEBUG nova.network.neutron [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Updating instance_info_cache with network_info: [{"id": "19fc1ad9-b48d-45ba-9f90-85b3a5910021", "address": "fa:16:3e:38:80:66", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fc1ad9-b4", "ovs_interfaceid": "19fc1ad9-b48d-45ba-9f90-85b3a5910021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.400912] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f11e3e-e4fe-43df-bc68-9ffded04ca7e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.412620] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ed4353-ed4b-479c-b274-c0c55f69429c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.444438] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-921c1f7e-39da-479c-a799-85826fece805 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.456505] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad3d54d-07f9-457d-8d2d-54cfef051bb1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.475371] env[62814]: DEBUG nova.compute.provider_tree [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.482900] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293605, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.585962} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.482900] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c9e11016-e92a-459e-b5ee-b0e43ce29450/c9e11016-e92a-459e-b5ee-b0e43ce29450.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 785.482900] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.482900] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-880496ff-2c24-4974-88f7-6988ccd1e17d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.489994] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 785.489994] env[62814]: value = "task-4293606" [ 785.489994] env[62814]: _type = "Task" [ 785.489994] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.498930] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293606, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.652762] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Releasing lock "refresh_cache-b289800f-49ce-49e7-b6bc-a3b4ec84b434" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 785.652762] env[62814]: DEBUG nova.compute.manager [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Instance network_info: |[{"id": "19fc1ad9-b48d-45ba-9f90-85b3a5910021", "address": "fa:16:3e:38:80:66", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fc1ad9-b4", "ovs_interfaceid": "19fc1ad9-b48d-45ba-9f90-85b3a5910021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 785.653963] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:80:66', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '19fc1ad9-b48d-45ba-9f90-85b3a5910021', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 785.663211] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Creating folder: Project (447bee1ea28948ebaa134d807e481edc). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.663550] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b080338d-60fd-402a-8f15-2f3e7dc4db97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.675801] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Created folder: Project (447bee1ea28948ebaa134d807e481edc) in parent group-v845547. [ 785.676106] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Creating folder: Instances. Parent ref: group-v845662. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 785.676425] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-47dcbf42-2cd5-45b5-8740-cf0bb4d31c74 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.687047] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Created folder: Instances in parent group-v845662. [ 785.687765] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 785.688037] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 785.688667] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be375a57-61b3-446a-958c-b3ec28026bd2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.716269] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 785.716269] env[62814]: value = "task-4293609" [ 785.716269] env[62814]: _type = "Task" [ 785.716269] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.726041] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293609, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.745575] env[62814]: DEBUG nova.compute.manager [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 785.777588] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 785.777932] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 785.778043] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 785.778229] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 785.778374] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 785.779218] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 785.779410] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 785.779532] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 785.779752] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 785.779906] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 785.780092] env[62814]: DEBUG nova.virt.hardware [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 785.781757] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd73fc9-1ae2-4c04-b2a4-9d0e666b3230 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.791965] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed011c9-af59-43fc-916a-5374b578ed14 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.885830] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "9683894b-a300-4400-a1b9-db62478f42c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.886122] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.925287] env[62814]: DEBUG nova.compute.manager [req-0d726e7d-e167-48f5-b439-c619084f3db8 req-b080011e-2473-442e-84bb-9d8b725b59ea service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Received event network-changed-19fc1ad9-b48d-45ba-9f90-85b3a5910021 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 785.925562] env[62814]: DEBUG nova.compute.manager [req-0d726e7d-e167-48f5-b439-c619084f3db8 req-b080011e-2473-442e-84bb-9d8b725b59ea service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Refreshing instance network info cache due to event network-changed-19fc1ad9-b48d-45ba-9f90-85b3a5910021. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 785.925716] env[62814]: DEBUG oslo_concurrency.lockutils [req-0d726e7d-e167-48f5-b439-c619084f3db8 req-b080011e-2473-442e-84bb-9d8b725b59ea service nova] Acquiring lock "refresh_cache-b289800f-49ce-49e7-b6bc-a3b4ec84b434" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.925853] env[62814]: DEBUG oslo_concurrency.lockutils [req-0d726e7d-e167-48f5-b439-c619084f3db8 req-b080011e-2473-442e-84bb-9d8b725b59ea service nova] Acquired lock "refresh_cache-b289800f-49ce-49e7-b6bc-a3b4ec84b434" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 785.926331] env[62814]: DEBUG nova.network.neutron [req-0d726e7d-e167-48f5-b439-c619084f3db8 req-b080011e-2473-442e-84bb-9d8b725b59ea service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Refreshing network info cache for port 19fc1ad9-b48d-45ba-9f90-85b3a5910021 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 785.950895] env[62814]: DEBUG nova.compute.manager [req-1bb4b5e5-9e45-4d13-83f6-05912d1c142f req-2c938f2f-24fe-4de0-b87c-3a148609b030 service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Received event network-vif-deleted-7aee2670-a9fd-4740-9463-7492279dcfc6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 785.951250] env[62814]: INFO nova.compute.manager [req-1bb4b5e5-9e45-4d13-83f6-05912d1c142f req-2c938f2f-24fe-4de0-b87c-3a148609b030 service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Neutron deleted interface 7aee2670-a9fd-4740-9463-7492279dcfc6; detaching it from the instance and deleting it from the info cache [ 785.951900] env[62814]: DEBUG nova.network.neutron [req-1bb4b5e5-9e45-4d13-83f6-05912d1c142f req-2c938f2f-24fe-4de0-b87c-3a148609b030 service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.979060] env[62814]: DEBUG nova.scheduler.client.report [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 785.999965] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293606, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080075} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.000255] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.001144] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb7f16a-dd1e-4d6b-b0dc-cf8aa36f80d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.025637] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] c9e11016-e92a-459e-b5ee-b0e43ce29450/c9e11016-e92a-459e-b5ee-b0e43ce29450.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.026563] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-987ae35f-b77f-4561-a594-3f4c0c48eb3c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.047643] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 786.047643] env[62814]: value = "task-4293610" [ 786.047643] env[62814]: _type = "Task" [ 786.047643] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.056189] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293610, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.226860] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293609, 'name': CreateVM_Task, 'duration_secs': 0.366658} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.227122] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 786.227807] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.227989] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.228789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 786.228789] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f99bbf06-028b-4ae1-9609-08804ecc5a68 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.233835] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for the task: (returnval){ [ 786.233835] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d1f20d-c9eb-42c9-2a61-ddc126a2290a" [ 786.233835] env[62814]: _type = "Task" [ 786.233835] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.241881] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d1f20d-c9eb-42c9-2a61-ddc126a2290a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.324542] env[62814]: DEBUG nova.network.neutron [-] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.421748] env[62814]: DEBUG nova.network.neutron [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Successfully updated port: 1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 786.457205] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a3222e6-a48a-49ec-ac17-e94af974aa8d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.466158] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a8f1ac7-b3da-4887-ba98-f802ff641165 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.484453] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.767s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 786.485095] env[62814]: DEBUG nova.compute.manager [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 786.487723] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 32.872s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 786.493189] env[62814]: INFO nova.compute.claims [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.513348] env[62814]: DEBUG nova.compute.manager [req-1bb4b5e5-9e45-4d13-83f6-05912d1c142f req-2c938f2f-24fe-4de0-b87c-3a148609b030 service nova] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Detach interface failed, port_id=7aee2670-a9fd-4740-9463-7492279dcfc6, reason: Instance 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 786.559236] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293610, 'name': ReconfigVM_Task, 'duration_secs': 0.357804} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.559949] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Reconfigured VM instance instance-00000027 to attach disk [datastore2] c9e11016-e92a-459e-b5ee-b0e43ce29450/c9e11016-e92a-459e-b5ee-b0e43ce29450.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.560754] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ce3ddab-00a1-4a25-88fc-6db727daec70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.567301] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 786.567301] env[62814]: value = "task-4293611" [ 786.567301] env[62814]: _type = "Task" [ 786.567301] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.576110] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293611, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.615190] env[62814]: DEBUG oslo_vmware.rw_handles [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ad171e-cac5-2a80-441e-698d3f02d9df/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 786.616116] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015cdc77-a815-4d08-8a82-78cfa7f5bc1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.623626] env[62814]: DEBUG oslo_vmware.rw_handles [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ad171e-cac5-2a80-441e-698d3f02d9df/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 786.623793] env[62814]: ERROR oslo_vmware.rw_handles [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ad171e-cac5-2a80-441e-698d3f02d9df/disk-0.vmdk due to incomplete transfer. [ 786.624039] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-6a82b7f8-54c7-4284-bf69-ce873b2f2603 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.632806] env[62814]: DEBUG oslo_vmware.rw_handles [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52ad171e-cac5-2a80-441e-698d3f02d9df/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 786.633013] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Uploaded image c73af40e-dd5c-4c62-8e27-a5a56963e1c7 to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 786.635063] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 786.636298] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-368a1843-8490-4a4a-9e91-3f2d15b59ac5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.640784] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 786.640784] env[62814]: value = "task-4293612" [ 786.640784] env[62814]: _type = "Task" [ 786.640784] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.648844] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293612, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.722733] env[62814]: DEBUG nova.network.neutron [req-0d726e7d-e167-48f5-b439-c619084f3db8 req-b080011e-2473-442e-84bb-9d8b725b59ea service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Updated VIF entry in instance network info cache for port 19fc1ad9-b48d-45ba-9f90-85b3a5910021. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 786.723167] env[62814]: DEBUG nova.network.neutron [req-0d726e7d-e167-48f5-b439-c619084f3db8 req-b080011e-2473-442e-84bb-9d8b725b59ea service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Updating instance_info_cache with network_info: [{"id": "19fc1ad9-b48d-45ba-9f90-85b3a5910021", "address": "fa:16:3e:38:80:66", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fc1ad9-b4", "ovs_interfaceid": "19fc1ad9-b48d-45ba-9f90-85b3a5910021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.746245] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d1f20d-c9eb-42c9-2a61-ddc126a2290a, 'name': SearchDatastore_Task, 'duration_secs': 0.018711} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.747785] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 786.747785] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 786.747785] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.747785] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.747785] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 786.747785] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-631cae27-4f38-4ef2-915f-34597cdce85e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.756959] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 786.757380] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 786.757999] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-483362fb-a80c-45c6-b4d3-c9b3a2bd6996 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.764633] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for the task: (returnval){ [ 786.764633] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d65d7f-60b2-e7cd-21c6-f4a6e2247f98" [ 786.764633] env[62814]: _type = "Task" [ 786.764633] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.773216] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d65d7f-60b2-e7cd-21c6-f4a6e2247f98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.827163] env[62814]: INFO nova.compute.manager [-] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Took 1.86 seconds to deallocate network for instance. [ 786.924825] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquiring lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.924985] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquired lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 786.925151] env[62814]: DEBUG nova.network.neutron [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 787.000301] env[62814]: DEBUG nova.compute.utils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 787.001697] env[62814]: DEBUG nova.compute.manager [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Not allocating networking since 'none' was specified. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 787.078918] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293611, 'name': Rename_Task, 'duration_secs': 0.316189} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.078918] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.078918] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35858ef2-fac6-45f5-8b3b-25d498b2ae5e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.085416] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 787.085416] env[62814]: value = "task-4293613" [ 787.085416] env[62814]: _type = "Task" [ 787.085416] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.093489] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293613, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.150126] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293612, 'name': Destroy_Task, 'duration_secs': 0.310556} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.150443] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Destroyed the VM [ 787.150770] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 787.151063] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ee7de183-b2a7-4402-b9af-378dc3c2eff0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.156714] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 787.156714] env[62814]: value = "task-4293614" [ 787.156714] env[62814]: _type = "Task" [ 787.156714] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.164396] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293614, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.229398] env[62814]: DEBUG oslo_concurrency.lockutils [req-0d726e7d-e167-48f5-b439-c619084f3db8 req-b080011e-2473-442e-84bb-9d8b725b59ea service nova] Releasing lock "refresh_cache-b289800f-49ce-49e7-b6bc-a3b4ec84b434" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.276231] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d65d7f-60b2-e7cd-21c6-f4a6e2247f98, 'name': SearchDatastore_Task, 'duration_secs': 0.016828} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.277141] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64b1fa32-f052-49a7-a722-ce45e2fa493b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.282170] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for the task: (returnval){ [ 787.282170] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52766c3e-7cdc-f812-5e3c-353e58fa94c7" [ 787.282170] env[62814]: _type = "Task" [ 787.282170] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.290062] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52766c3e-7cdc-f812-5e3c-353e58fa94c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.334282] env[62814]: DEBUG oslo_concurrency.lockutils [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.462715] env[62814]: DEBUG nova.network.neutron [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.502946] env[62814]: DEBUG nova.compute.manager [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 787.596528] env[62814]: DEBUG oslo_vmware.api [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293613, 'name': PowerOnVM_Task, 'duration_secs': 0.482505} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.598945] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 787.599239] env[62814]: DEBUG nova.compute.manager [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 787.600264] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d0bece-99cb-4768-9645-dfbff9e750fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.648924] env[62814]: DEBUG nova.network.neutron [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updating instance_info_cache with network_info: [{"id": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "address": "fa:16:3e:8e:bc:5f", "network": {"id": "ddc113f9-9fc4-4847-9181-82a3ee75ec5c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-787311776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6254d7e15b4c5ba267e8da03d1b82d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1865cb89-88", "ovs_interfaceid": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.667314] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293614, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.792979] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52766c3e-7cdc-f812-5e3c-353e58fa94c7, 'name': SearchDatastore_Task, 'duration_secs': 0.010215} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.795592] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 787.795854] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b289800f-49ce-49e7-b6bc-a3b4ec84b434/b289800f-49ce-49e7-b6bc-a3b4ec84b434.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 787.796454] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d793bd3c-8dcd-4c17-8939-9c4ab2e22597 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.804645] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for the task: (returnval){ [ 787.804645] env[62814]: value = "task-4293615" [ 787.804645] env[62814]: _type = "Task" [ 787.804645] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.816557] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293615, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.001507] env[62814]: DEBUG nova.compute.manager [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Received event network-vif-plugged-1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 788.001710] env[62814]: DEBUG oslo_concurrency.lockutils [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] Acquiring lock "d366a755-49b4-427b-8564-d8572a7fbbb7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.001988] env[62814]: DEBUG oslo_concurrency.lockutils [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] Lock "d366a755-49b4-427b-8564-d8572a7fbbb7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.002293] env[62814]: DEBUG oslo_concurrency.lockutils [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] Lock "d366a755-49b4-427b-8564-d8572a7fbbb7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.002580] env[62814]: DEBUG nova.compute.manager [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] No waiting events found dispatching network-vif-plugged-1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 788.002863] env[62814]: WARNING nova.compute.manager [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Received unexpected event network-vif-plugged-1865cb89-88d6-4e84-81ae-daece6adaa4d for instance with vm_state building and task_state spawning. [ 788.003156] env[62814]: DEBUG nova.compute.manager [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Received event network-changed-1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 788.003460] env[62814]: DEBUG nova.compute.manager [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Refreshing instance network info cache due to event network-changed-1865cb89-88d6-4e84-81ae-daece6adaa4d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 788.003776] env[62814]: DEBUG oslo_concurrency.lockutils [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] Acquiring lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.093262] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db293b5d-27a7-4670-abeb-b527ba5f53fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.103046] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7feecae7-ed02-4b4a-b104-1222953b24c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.159036] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Releasing lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 788.159036] env[62814]: DEBUG nova.compute.manager [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Instance network_info: |[{"id": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "address": "fa:16:3e:8e:bc:5f", "network": {"id": "ddc113f9-9fc4-4847-9181-82a3ee75ec5c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-787311776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6254d7e15b4c5ba267e8da03d1b82d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1865cb89-88", "ovs_interfaceid": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 788.159036] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50761c3-5244-4218-91a8-792556490e1d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.163396] env[62814]: DEBUG oslo_concurrency.lockutils [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.165867] env[62814]: DEBUG oslo_concurrency.lockutils [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] Acquired lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 788.165867] env[62814]: DEBUG nova.network.neutron [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Refreshing network info cache for port 1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 788.165867] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:bc:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c47e98ff-83cf-48d2-bf91-2931c7386b6a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1865cb89-88d6-4e84-81ae-daece6adaa4d', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.173443] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Creating folder: Project (fd6254d7e15b4c5ba267e8da03d1b82d). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 788.178296] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-29b14afe-a9e2-4798-a342-623b685a0b4a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.187751] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7755f1a-b753-48c4-be5a-c5a95c908847 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.192579] env[62814]: DEBUG oslo_vmware.api [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293614, 'name': RemoveSnapshot_Task, 'duration_secs': 0.619164} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.194246] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 788.194520] env[62814]: INFO nova.compute.manager [None req-8bf4c76b-28f9-4895-b79e-f5ea4289194b tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Took 14.15 seconds to snapshot the instance on the hypervisor. [ 788.197103] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Created folder: Project (fd6254d7e15b4c5ba267e8da03d1b82d) in parent group-v845547. [ 788.197236] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Creating folder: Instances. Parent ref: group-v845665. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 788.198455] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a4c20df-9933-4f55-a5cc-7601aff5dc40 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.209041] env[62814]: DEBUG nova.compute.provider_tree [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.217634] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Created folder: Instances in parent group-v845665. [ 788.217917] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 788.218161] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 788.218385] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-120f8a23-7d69-4f86-a0b6-b2af0f53067f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.238422] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.238422] env[62814]: value = "task-4293618" [ 788.238422] env[62814]: _type = "Task" [ 788.238422] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.247147] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293618, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.314874] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293615, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.349158] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "c9e11016-e92a-459e-b5ee-b0e43ce29450" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.349512] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "c9e11016-e92a-459e-b5ee-b0e43ce29450" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.349972] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "c9e11016-e92a-459e-b5ee-b0e43ce29450-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.350097] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "c9e11016-e92a-459e-b5ee-b0e43ce29450-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.350329] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "c9e11016-e92a-459e-b5ee-b0e43ce29450-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.352996] env[62814]: INFO nova.compute.manager [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Terminating instance [ 788.515579] env[62814]: DEBUG nova.compute.manager [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 788.541172] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 788.541421] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.541575] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 788.541776] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.541968] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 788.542132] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 788.542341] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 788.542499] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 788.542662] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 788.542827] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 788.542994] env[62814]: DEBUG nova.virt.hardware [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 788.544177] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b13a090-eb30-4a6c-b5ff-85330b9bdd23 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.552334] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e91e3c1-31bd-41ec-a98a-d8b3caee02a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.565763] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 788.571497] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Creating folder: Project (42d432b504604f10b5f4cbb3b82452ae). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 788.571547] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f929643-4cb4-4db2-97a0-011507f67782 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.582065] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Created folder: Project (42d432b504604f10b5f4cbb3b82452ae) in parent group-v845547. [ 788.582276] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Creating folder: Instances. Parent ref: group-v845668. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 788.582492] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09679765-83f2-46d9-b8eb-9eb447528658 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.591641] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Created folder: Instances in parent group-v845668. [ 788.591861] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 788.592065] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 788.592254] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52a33cd0-b607-408f-92b4-a630d4423cd9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.607699] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 788.607699] env[62814]: value = "task-4293621" [ 788.607699] env[62814]: _type = "Task" [ 788.607699] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.615091] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293621, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.712251] env[62814]: DEBUG nova.scheduler.client.report [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 788.748838] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293618, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.816134] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293615, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.654159} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.816134] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b289800f-49ce-49e7-b6bc-a3b4ec84b434/b289800f-49ce-49e7-b6bc-a3b4ec84b434.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 788.816389] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 788.816553] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3e7740ac-d678-4f98-82a4-e30741d711f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.823449] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for the task: (returnval){ [ 788.823449] env[62814]: value = "task-4293622" [ 788.823449] env[62814]: _type = "Task" [ 788.823449] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.832931] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.857982] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "refresh_cache-c9e11016-e92a-459e-b5ee-b0e43ce29450" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.858280] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquired lock "refresh_cache-c9e11016-e92a-459e-b5ee-b0e43ce29450" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 788.858564] env[62814]: DEBUG nova.network.neutron [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.933883] env[62814]: DEBUG nova.network.neutron [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updated VIF entry in instance network info cache for port 1865cb89-88d6-4e84-81ae-daece6adaa4d. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 788.934271] env[62814]: DEBUG nova.network.neutron [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updating instance_info_cache with network_info: [{"id": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "address": "fa:16:3e:8e:bc:5f", "network": {"id": "ddc113f9-9fc4-4847-9181-82a3ee75ec5c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-787311776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6254d7e15b4c5ba267e8da03d1b82d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1865cb89-88", "ovs_interfaceid": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.118320] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293621, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.217451] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.730s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.218050] env[62814]: DEBUG nova.compute.manager [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 789.221172] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.470s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 789.221373] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.223517] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.158s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 789.224919] env[62814]: INFO nova.compute.claims [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.250875] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293618, 'name': CreateVM_Task, 'duration_secs': 0.7273} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.251644] env[62814]: INFO nova.scheduler.client.report [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Deleted allocations for instance a589a3d8-20dc-4ff5-a192-c540e29f39d6 [ 789.252537] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.253764] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.253764] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.253764] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 789.256163] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76a3e1df-da3b-4c74-8264-2bb0c6118988 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.262575] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 789.262575] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529a0065-a21c-3fc1-9928-2c764eaca001" [ 789.262575] env[62814]: _type = "Task" [ 789.262575] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.270529] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529a0065-a21c-3fc1-9928-2c764eaca001, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.332771] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074349} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.333051] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 789.333854] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe79024-0940-452f-8b9d-e761387ddfc9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.355998] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] b289800f-49ce-49e7-b6bc-a3b4ec84b434/b289800f-49ce-49e7-b6bc-a3b4ec84b434.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 789.356640] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e70dc29d-8812-47a4-ad8d-47579eab1ea8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.378033] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for the task: (returnval){ [ 789.378033] env[62814]: value = "task-4293623" [ 789.378033] env[62814]: _type = "Task" [ 789.378033] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.386245] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293623, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.391189] env[62814]: DEBUG nova.network.neutron [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.436857] env[62814]: DEBUG oslo_concurrency.lockutils [req-72c0210f-e443-456e-9c30-99c64de8c2da req-b2509be2-5563-4d31-90aa-58d00658e8e0 service nova] Releasing lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.473795] env[62814]: DEBUG nova.network.neutron [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.620102] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293621, 'name': CreateVM_Task, 'duration_secs': 0.622852} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.620658] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 789.620939] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.731505] env[62814]: DEBUG nova.compute.utils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 789.735161] env[62814]: DEBUG nova.compute.manager [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 789.736140] env[62814]: DEBUG nova.network.neutron [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 789.761300] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0981304f-3a77-4507-8473-b6c7f6830d97 tempest-ServerMetadataTestJSON-1096886146 tempest-ServerMetadataTestJSON-1096886146-project-member] Lock "a589a3d8-20dc-4ff5-a192-c540e29f39d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.624s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 789.772540] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529a0065-a21c-3fc1-9928-2c764eaca001, 'name': SearchDatastore_Task, 'duration_secs': 0.009667} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.772828] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.773193] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 789.773298] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.773489] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.773616] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 789.773889] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.774443] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 789.774443] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3b2d2de8-70bc-4901-b95b-9ce83d6c0d07 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.776276] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9363b2d2-3a03-4958-ad0c-9705c142da52 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.784490] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 789.784490] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52444e3c-9a99-f1dd-f235-fe46c1aa7570" [ 789.784490] env[62814]: _type = "Task" [ 789.784490] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.788482] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 789.788847] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 789.790133] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66c444cd-1bea-4e27-93e0-53dee1a47209 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.801231] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52444e3c-9a99-f1dd-f235-fe46c1aa7570, 'name': SearchDatastore_Task, 'duration_secs': 0.010107} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.801231] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.801231] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 789.801630] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.802853] env[62814]: DEBUG nova.policy [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '307242d782d24d1594ee89c0e46e7aa3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42a09e557c794726aa4895c1563ea62f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 789.807634] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 789.807634] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52609250-3ced-8562-aed9-098ffd740eb1" [ 789.807634] env[62814]: _type = "Task" [ 789.807634] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.823041] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52609250-3ced-8562-aed9-098ffd740eb1, 'name': SearchDatastore_Task, 'duration_secs': 0.00855} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.823949] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ae95023-6fe8-4d15-abaf-82fb99d14264 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.830886] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 789.830886] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd2435-cdf4-5819-3bb8-899986a38ef4" [ 789.830886] env[62814]: _type = "Task" [ 789.830886] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.842494] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd2435-cdf4-5819-3bb8-899986a38ef4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.887231] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293623, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.979118] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Releasing lock "refresh_cache-c9e11016-e92a-459e-b5ee-b0e43ce29450" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.979118] env[62814]: DEBUG nova.compute.manager [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 789.979118] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.979118] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f64720c-e81d-4188-bb42-d9bdaec8a1f3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.988071] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 789.988071] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a09bba98-b992-4b07-80cd-8cc1bb5c3b68 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.993925] env[62814]: DEBUG oslo_vmware.api [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 789.993925] env[62814]: value = "task-4293624" [ 789.993925] env[62814]: _type = "Task" [ 789.993925] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.002778] env[62814]: DEBUG oslo_vmware.api [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.240195] env[62814]: DEBUG nova.compute.manager [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 790.272690] env[62814]: DEBUG nova.network.neutron [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Successfully created port: 95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 790.343317] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd2435-cdf4-5819-3bb8-899986a38ef4, 'name': SearchDatastore_Task, 'duration_secs': 0.008888} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.346365] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.346840] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d366a755-49b4-427b-8564-d8572a7fbbb7/d366a755-49b4-427b-8564-d8572a7fbbb7.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 790.348040] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 790.348040] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.348040] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f58bc33f-4dfb-4c28-8656-ade84ffd06c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.350283] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eeb34089-ad61-44ed-963b-be04f75cccdf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.362020] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 790.362020] env[62814]: value = "task-4293625" [ 790.362020] env[62814]: _type = "Task" [ 790.362020] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.362020] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.362020] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 790.365479] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-21725cec-bc92-4aa3-a3e9-f29dae406089 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.376833] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293625, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.378333] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 790.378333] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52647799-d136-a901-6b02-a211d48cba48" [ 790.378333] env[62814]: _type = "Task" [ 790.378333] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.390134] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52647799-d136-a901-6b02-a211d48cba48, 'name': SearchDatastore_Task, 'duration_secs': 0.009387} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.396526] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293623, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.397021] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-931d6500-2a0e-44a0-a94a-4c3aa2273f4c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.402250] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 790.402250] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e0f110-4ec2-6cee-450f-aa208a9efdd8" [ 790.402250] env[62814]: _type = "Task" [ 790.402250] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.410269] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e0f110-4ec2-6cee-450f-aa208a9efdd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.506958] env[62814]: DEBUG oslo_vmware.api [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.876732] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293625, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511197} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.877122] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d366a755-49b4-427b-8564-d8572a7fbbb7/d366a755-49b4-427b-8564-d8572a7fbbb7.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 790.877417] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 790.877758] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e9ad329b-c83e-4307-8769-b6aff6b9e593 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.889961] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 790.889961] env[62814]: value = "task-4293626" [ 790.889961] env[62814]: _type = "Task" [ 790.889961] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.898416] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293623, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.902445] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293626, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.912167] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e0f110-4ec2-6cee-450f-aa208a9efdd8, 'name': SearchDatastore_Task, 'duration_secs': 0.008104} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.913371] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.913647] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 127fee64-fd56-4a23-bdd2-18c817898fd5/127fee64-fd56-4a23-bdd2-18c817898fd5.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 790.914125] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-884fd6db-c8e3-4c45-a235-fdbc06855e5a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.920210] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 790.920210] env[62814]: value = "task-4293627" [ 790.920210] env[62814]: _type = "Task" [ 790.920210] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.925649] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d58ec18-9719-4193-8581-6e04dbf9465e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.931403] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293627, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.936490] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ff4ff3-c181-4861-b8d5-c3a3abed0ce2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.970938] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ded7ef-616e-45da-94e5-5b49a4d8cbbd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.980587] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2baa167-df68-44ca-a72f-f683399ae41e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.004712] env[62814]: DEBUG nova.compute.provider_tree [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.019591] env[62814]: DEBUG oslo_vmware.api [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.179198] env[62814]: DEBUG nova.compute.manager [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 791.179198] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0a8a75-8c3f-402a-828e-fefefd56bdb1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.256581] env[62814]: DEBUG nova.compute.manager [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 791.289892] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 791.289892] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.289892] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 791.289892] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.289892] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 791.289892] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 791.289892] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 791.289892] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 791.290251] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 791.290369] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 791.290534] env[62814]: DEBUG nova.virt.hardware [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 791.291602] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433e4d67-b685-4a7b-aa84-4770ef9c7a2d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.300491] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dfd3270-81d3-4df8-a43d-bb543e238a57 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.389455] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293623, 'name': ReconfigVM_Task, 'duration_secs': 1.622879} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.389820] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Reconfigured VM instance instance-00000028 to attach disk [datastore2] b289800f-49ce-49e7-b6bc-a3b4ec84b434/b289800f-49ce-49e7-b6bc-a3b4ec84b434.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.390534] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c98b1eca-5f52-46e2-9205-3cfc1f111b35 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.400600] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293626, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.130978} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.402035] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 791.402715] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for the task: (returnval){ [ 791.402715] env[62814]: value = "task-4293628" [ 791.402715] env[62814]: _type = "Task" [ 791.402715] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.403509] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76bde114-6c7c-416e-a9b9-2b1ca0eae1d8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.431383] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Reconfiguring VM instance instance-00000029 to attach disk [datastore2] d366a755-49b4-427b-8564-d8572a7fbbb7/d366a755-49b4-427b-8564-d8572a7fbbb7.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 791.439971] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc01a5ec-6c1c-4c48-9919-1981ac564b90 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.456952] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293628, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.462709] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293627, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481142} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.468019] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 127fee64-fd56-4a23-bdd2-18c817898fd5/127fee64-fd56-4a23-bdd2-18c817898fd5.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 791.468019] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.468019] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 791.468019] env[62814]: value = "task-4293629" [ 791.468019] env[62814]: _type = "Task" [ 791.468019] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.468019] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0403d065-0823-47e8-a950-0a3ef04ce789 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.477393] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293629, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.478864] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 791.478864] env[62814]: value = "task-4293630" [ 791.478864] env[62814]: _type = "Task" [ 791.478864] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.486894] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293630, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.509969] env[62814]: DEBUG nova.scheduler.client.report [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 791.519176] env[62814]: DEBUG oslo_vmware.api [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293624, 'name': PowerOffVM_Task, 'duration_secs': 1.091035} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.519427] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 791.519660] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 791.519958] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae13238f-9665-4c00-865b-064988cac43a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.544750] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 791.545049] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 791.545207] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Deleting the datastore file [datastore2] c9e11016-e92a-459e-b5ee-b0e43ce29450 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 791.545505] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f0a0a07a-3500-4f60-b175-da4c557eef2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.552655] env[62814]: DEBUG oslo_vmware.api [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for the task: (returnval){ [ 791.552655] env[62814]: value = "task-4293632" [ 791.552655] env[62814]: _type = "Task" [ 791.552655] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.560921] env[62814]: DEBUG oslo_vmware.api [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293632, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.692463] env[62814]: INFO nova.compute.manager [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] instance snapshotting [ 791.695156] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eaf1d5e-291e-4a9f-865e-b5cf91c66dfb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.716827] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1e017f-2f06-4850-95fe-76a27f509192 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.917024] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293628, 'name': Rename_Task, 'duration_secs': 0.247331} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.917024] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 791.917449] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d336f5bd-1c32-4650-b6ee-c216b67d0bb4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.923400] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for the task: (returnval){ [ 791.923400] env[62814]: value = "task-4293633" [ 791.923400] env[62814]: _type = "Task" [ 791.923400] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.934704] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293633, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.977862] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293629, 'name': ReconfigVM_Task, 'duration_secs': 0.397276} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.978592] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Reconfigured VM instance instance-00000029 to attach disk [datastore2] d366a755-49b4-427b-8564-d8572a7fbbb7/d366a755-49b4-427b-8564-d8572a7fbbb7.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 791.979373] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d4213b13-388c-485b-88c7-d5953e0e3e43 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.988155] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293630, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112318} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.989384] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 791.989742] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 791.989742] env[62814]: value = "task-4293634" [ 791.989742] env[62814]: _type = "Task" [ 791.989742] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.990413] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ecc3b5a-30c5-4182-b61c-e37811c66732 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.012646] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 127fee64-fd56-4a23-bdd2-18c817898fd5/127fee64-fd56-4a23-bdd2-18c817898fd5.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.016042] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d32180d7-9119-46c9-9a19-e88edcb99344 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.030636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.807s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.031152] env[62814]: DEBUG nova.compute.manager [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 792.033639] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293634, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.034379] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.584s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.034569] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.036465] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.629s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.038419] env[62814]: INFO nova.compute.claims [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.046574] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 792.046574] env[62814]: value = "task-4293635" [ 792.046574] env[62814]: _type = "Task" [ 792.046574] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.057967] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293635, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.065159] env[62814]: DEBUG oslo_vmware.api [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Task: {'id': task-4293632, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100875} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.065443] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 792.065604] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 792.065782] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 792.065961] env[62814]: INFO nova.compute.manager [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Took 2.09 seconds to destroy the instance on the hypervisor. [ 792.066226] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 792.066412] env[62814]: DEBUG nova.compute.manager [-] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 792.066505] env[62814]: DEBUG nova.network.neutron [-] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 792.070470] env[62814]: INFO nova.scheduler.client.report [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Deleted allocations for instance 369cd937-4c18-4068-ae59-70a1d585094b [ 792.093851] env[62814]: DEBUG nova.network.neutron [-] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.228148] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 792.228475] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-14a4949f-060b-455d-b383-4e6457e070ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.240379] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 792.240379] env[62814]: value = "task-4293636" [ 792.240379] env[62814]: _type = "Task" [ 792.240379] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.248632] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293636, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.304731] env[62814]: DEBUG nova.compute.manager [req-f4fa0b33-9196-448b-9b8f-cfe82a852f4d req-80b920b7-be55-4ac0-80b9-6cf848f296f8 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Received event network-vif-plugged-95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 792.304731] env[62814]: DEBUG oslo_concurrency.lockutils [req-f4fa0b33-9196-448b-9b8f-cfe82a852f4d req-80b920b7-be55-4ac0-80b9-6cf848f296f8 service nova] Acquiring lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.305727] env[62814]: DEBUG oslo_concurrency.lockutils [req-f4fa0b33-9196-448b-9b8f-cfe82a852f4d req-80b920b7-be55-4ac0-80b9-6cf848f296f8 service nova] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.305727] env[62814]: DEBUG oslo_concurrency.lockutils [req-f4fa0b33-9196-448b-9b8f-cfe82a852f4d req-80b920b7-be55-4ac0-80b9-6cf848f296f8 service nova] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.305727] env[62814]: DEBUG nova.compute.manager [req-f4fa0b33-9196-448b-9b8f-cfe82a852f4d req-80b920b7-be55-4ac0-80b9-6cf848f296f8 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] No waiting events found dispatching network-vif-plugged-95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 792.305727] env[62814]: WARNING nova.compute.manager [req-f4fa0b33-9196-448b-9b8f-cfe82a852f4d req-80b920b7-be55-4ac0-80b9-6cf848f296f8 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Received unexpected event network-vif-plugged-95b486fb-bbe0-4666-a926-01bcfa37eaad for instance with vm_state building and task_state spawning. [ 792.432971] env[62814]: DEBUG nova.network.neutron [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Successfully updated port: 95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 792.442138] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293633, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.505350] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293634, 'name': Rename_Task, 'duration_secs': 0.200834} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.505696] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 792.506080] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-128284d5-724d-4742-be34-dc625999fe62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.513898] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 792.513898] env[62814]: value = "task-4293637" [ 792.513898] env[62814]: _type = "Task" [ 792.513898] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.527258] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293637, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.537166] env[62814]: DEBUG nova.compute.utils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 792.539131] env[62814]: DEBUG nova.compute.manager [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 792.539469] env[62814]: DEBUG nova.network.neutron [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 792.561338] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293635, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.583433] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174b485c-1a91-46f1-8396-7457dfb9ba2e tempest-FloatingIPsAssociationNegativeTestJSON-1005726298 tempest-FloatingIPsAssociationNegativeTestJSON-1005726298-project-member] Lock "369cd937-4c18-4068-ae59-70a1d585094b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.150s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.596035] env[62814]: DEBUG nova.network.neutron [-] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.623710] env[62814]: DEBUG nova.policy [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '713f3b7e15404bffa24349dac34c8534', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f11d1ef620764fa4b1e2b718ac207a44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 792.752307] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293636, 'name': CreateSnapshot_Task, 'duration_secs': 0.507162} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.753794] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 792.753794] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9a5af2-c58f-406b-88c7-c101f5c73778 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.939697] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.939913] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 792.940090] env[62814]: DEBUG nova.network.neutron [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.942117] env[62814]: DEBUG oslo_vmware.api [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293633, 'name': PowerOnVM_Task, 'duration_secs': 0.77244} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.942399] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 792.942945] env[62814]: INFO nova.compute.manager [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Took 9.86 seconds to spawn the instance on the hypervisor. [ 792.942945] env[62814]: DEBUG nova.compute.manager [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 792.944101] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e30af16-d2f0-4370-a670-af27f07a3374 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.025158] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293637, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.040180] env[62814]: DEBUG nova.compute.manager [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 793.060143] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293635, 'name': ReconfigVM_Task, 'duration_secs': 0.538233} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.060143] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 127fee64-fd56-4a23-bdd2-18c817898fd5/127fee64-fd56-4a23-bdd2-18c817898fd5.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.060566] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5c8b7b4-853e-41f8-99e6-34cd56ce04c2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.068072] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 793.068072] env[62814]: value = "task-4293638" [ 793.068072] env[62814]: _type = "Task" [ 793.068072] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.077613] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293638, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.101036] env[62814]: INFO nova.compute.manager [-] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Took 1.03 seconds to deallocate network for instance. [ 793.178273] env[62814]: DEBUG nova.network.neutron [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Successfully created port: 1b565819-b3b3-4e33-b334-07d6fd8b9d62 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 793.272425] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 793.275284] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f1601c32-63f1-4be4-be8d-52abcd9ca39b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.284513] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 793.284513] env[62814]: value = "task-4293639" [ 793.284513] env[62814]: _type = "Task" [ 793.284513] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.296126] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293639, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.473908] env[62814]: INFO nova.compute.manager [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Took 48.28 seconds to build instance. [ 793.514247] env[62814]: DEBUG nova.network.neutron [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.528807] env[62814]: DEBUG oslo_vmware.api [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4293637, 'name': PowerOnVM_Task, 'duration_secs': 0.734019} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.529103] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 793.529458] env[62814]: INFO nova.compute.manager [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Took 7.78 seconds to spawn the instance on the hypervisor. [ 793.529669] env[62814]: DEBUG nova.compute.manager [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 793.531196] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad91629e-83fa-4a80-9809-1026263f5e9e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.584020] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293638, 'name': Rename_Task, 'duration_secs': 0.214876} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.584020] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 793.584020] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4781e99e-cafd-4be0-b7ec-d70bf851ab48 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.592402] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 793.592402] env[62814]: value = "task-4293640" [ 793.592402] env[62814]: _type = "Task" [ 793.592402] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.599595] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293640, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.611667] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.696122] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9aabfc2-e2d0-499a-ac74-7cf8eab29a31 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.704170] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b14f29a-f58b-41d9-9deb-232984791ee9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.742295] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f7958c-e095-4567-8c70-36f5fa1d2d48 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.750428] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35feda8-015b-4baf-81c8-04b41ea58c0c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.765020] env[62814]: DEBUG nova.compute.provider_tree [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 793.797375] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293639, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.850409] env[62814]: DEBUG nova.network.neutron [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updating instance_info_cache with network_info: [{"id": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "address": "fa:16:3e:aa:55:02", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b486fb-bb", "ovs_interfaceid": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.979218] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ca8e1d6-959b-4d26-bd3c-3406d0282b86 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.300s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 794.053014] env[62814]: INFO nova.compute.manager [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Took 44.74 seconds to build instance. [ 794.056121] env[62814]: DEBUG nova.compute.manager [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 794.105943] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293640, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.108366] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 794.108808] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 794.108808] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 794.108955] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 794.109064] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 794.109224] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 794.109435] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 794.109590] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 794.109787] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 794.109952] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 794.110397] env[62814]: DEBUG nova.virt.hardware [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 794.111311] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c4d19e-4f67-4e30-8a92-a5e70fe0cf69 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.120747] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db1a70c-a4d8-4b4f-ba7d-78e448644f27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.290955] env[62814]: ERROR nova.scheduler.client.report [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [req-51ba0da8-af87-4154-a106-0e666c4723b5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-51ba0da8-af87-4154-a106-0e666c4723b5"}]} [ 794.299101] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293639, 'name': CloneVM_Task} progress is 95%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.313587] env[62814]: DEBUG nova.compute.manager [None req-c8a3df2f-497a-406d-b857-9187558f5111 tempest-ServerExternalEventsTest-781461774 tempest-ServerExternalEventsTest-781461774-project] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Received event network-changed {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 794.314016] env[62814]: DEBUG nova.compute.manager [None req-c8a3df2f-497a-406d-b857-9187558f5111 tempest-ServerExternalEventsTest-781461774 tempest-ServerExternalEventsTest-781461774-project] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Refreshing instance network info cache due to event network-changed. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 794.315169] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8a3df2f-497a-406d-b857-9187558f5111 tempest-ServerExternalEventsTest-781461774 tempest-ServerExternalEventsTest-781461774-project] Acquiring lock "refresh_cache-b289800f-49ce-49e7-b6bc-a3b4ec84b434" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.315169] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8a3df2f-497a-406d-b857-9187558f5111 tempest-ServerExternalEventsTest-781461774 tempest-ServerExternalEventsTest-781461774-project] Acquired lock "refresh_cache-b289800f-49ce-49e7-b6bc-a3b4ec84b434" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.315169] env[62814]: DEBUG nova.network.neutron [None req-c8a3df2f-497a-406d-b857-9187558f5111 tempest-ServerExternalEventsTest-781461774 tempest-ServerExternalEventsTest-781461774-project] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 794.318586] env[62814]: DEBUG nova.scheduler.client.report [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 794.339227] env[62814]: DEBUG nova.scheduler.client.report [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 794.339687] env[62814]: DEBUG nova.compute.provider_tree [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 794.352824] env[62814]: DEBUG nova.scheduler.client.report [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 794.355023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 794.355353] env[62814]: DEBUG nova.compute.manager [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Instance network_info: |[{"id": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "address": "fa:16:3e:aa:55:02", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b486fb-bb", "ovs_interfaceid": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 794.356386] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:55:02', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9c621a9c-66f5-426a-8aab-bd8b2e912106', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '95b486fb-bbe0-4666-a926-01bcfa37eaad', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 794.366203] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 794.366539] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 794.367413] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1113b51-d8f7-4406-9c64-b4df921b1608 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.386054] env[62814]: DEBUG nova.scheduler.client.report [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 794.395923] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 794.395923] env[62814]: value = "task-4293641" [ 794.395923] env[62814]: _type = "Task" [ 794.395923] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.406201] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293641, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.430084] env[62814]: DEBUG nova.compute.manager [req-69cc1953-0f59-4f5c-a4fa-cb10cb1f4f80 req-1b005d2f-661a-4686-bc93-6b8fdac25550 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Received event network-changed-95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 794.430380] env[62814]: DEBUG nova.compute.manager [req-69cc1953-0f59-4f5c-a4fa-cb10cb1f4f80 req-1b005d2f-661a-4686-bc93-6b8fdac25550 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Refreshing instance network info cache due to event network-changed-95b486fb-bbe0-4666-a926-01bcfa37eaad. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 794.430808] env[62814]: DEBUG oslo_concurrency.lockutils [req-69cc1953-0f59-4f5c-a4fa-cb10cb1f4f80 req-1b005d2f-661a-4686-bc93-6b8fdac25550 service nova] Acquiring lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.431042] env[62814]: DEBUG oslo_concurrency.lockutils [req-69cc1953-0f59-4f5c-a4fa-cb10cb1f4f80 req-1b005d2f-661a-4686-bc93-6b8fdac25550 service nova] Acquired lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.431267] env[62814]: DEBUG nova.network.neutron [req-69cc1953-0f59-4f5c-a4fa-cb10cb1f4f80 req-1b005d2f-661a-4686-bc93-6b8fdac25550 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Refreshing network info cache for port 95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 794.484596] env[62814]: DEBUG nova.compute.manager [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 794.559203] env[62814]: DEBUG oslo_concurrency.lockutils [None req-673b6244-2f7b-4572-af5d-52dcfacb1ea8 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lock "d366a755-49b4-427b-8564-d8572a7fbbb7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.658s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 794.606332] env[62814]: DEBUG oslo_vmware.api [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293640, 'name': PowerOnVM_Task, 'duration_secs': 0.603205} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.607747] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 794.608152] env[62814]: INFO nova.compute.manager [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Took 6.09 seconds to spawn the instance on the hypervisor. [ 794.608457] env[62814]: DEBUG nova.compute.manager [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 794.610391] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a202e9-02c2-44eb-af0b-6d0469e309e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.800189] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293639, 'name': CloneVM_Task, 'duration_secs': 1.295416} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.800457] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Created linked-clone VM from snapshot [ 794.801221] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5620a232-8774-4a07-9662-43035dd10840 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.810288] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Uploading image 1a74e7e6-0277-48fd-9c78-c75a4976ef3e {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 794.848291] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 794.848291] env[62814]: value = "vm-845672" [ 794.848291] env[62814]: _type = "VirtualMachine" [ 794.848291] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 794.848543] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-adcc7bf4-818d-4c1e-85f2-39c2beaa8262 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.855862] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lease: (returnval){ [ 794.855862] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52780c0e-d921-af2b-2a6e-3dc676767c8a" [ 794.855862] env[62814]: _type = "HttpNfcLease" [ 794.855862] env[62814]: } obtained for exporting VM: (result){ [ 794.855862] env[62814]: value = "vm-845672" [ 794.855862] env[62814]: _type = "VirtualMachine" [ 794.855862] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 794.856136] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the lease: (returnval){ [ 794.856136] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52780c0e-d921-af2b-2a6e-3dc676767c8a" [ 794.856136] env[62814]: _type = "HttpNfcLease" [ 794.856136] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 794.863910] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 794.863910] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52780c0e-d921-af2b-2a6e-3dc676767c8a" [ 794.863910] env[62814]: _type = "HttpNfcLease" [ 794.863910] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 794.908085] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293641, 'name': CreateVM_Task, 'duration_secs': 0.329354} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.908293] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 794.911661] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.911661] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 794.911928] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 794.912535] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9998b14d-b935-4b37-b918-5cf304de23aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.919015] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 794.919015] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c30d9d-822b-247e-3ad1-8081aa213f9b" [ 794.919015] env[62814]: _type = "Task" [ 794.919015] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.925409] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c30d9d-822b-247e-3ad1-8081aa213f9b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.012640] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.028731] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquiring lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.029413] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 795.029712] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquiring lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.029935] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 795.030775] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 795.036820] env[62814]: INFO nova.compute.manager [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Terminating instance [ 795.041761] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb7d3b3-3507-4c31-b344-399253626858 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.054908] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2c3400-51c0-4b4f-a9c1-ccfcc684be18 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.061505] env[62814]: DEBUG nova.compute.manager [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 795.101309] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd1eb044-56b7-44b9-b056-b5aa2d2a9a05 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.108880] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7dd19a1-437d-49dd-ac72-8314e8d36483 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.118240] env[62814]: DEBUG nova.network.neutron [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Successfully updated port: 1b565819-b3b3-4e33-b334-07d6fd8b9d62 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 795.140188] env[62814]: DEBUG nova.compute.provider_tree [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 795.144652] env[62814]: INFO nova.compute.manager [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Took 42.80 seconds to build instance. [ 795.296507] env[62814]: DEBUG nova.network.neutron [None req-c8a3df2f-497a-406d-b857-9187558f5111 tempest-ServerExternalEventsTest-781461774 tempest-ServerExternalEventsTest-781461774-project] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Updating instance_info_cache with network_info: [{"id": "19fc1ad9-b48d-45ba-9f90-85b3a5910021", "address": "fa:16:3e:38:80:66", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap19fc1ad9-b4", "ovs_interfaceid": "19fc1ad9-b48d-45ba-9f90-85b3a5910021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.345322] env[62814]: DEBUG nova.network.neutron [req-69cc1953-0f59-4f5c-a4fa-cb10cb1f4f80 req-1b005d2f-661a-4686-bc93-6b8fdac25550 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updated VIF entry in instance network info cache for port 95b486fb-bbe0-4666-a926-01bcfa37eaad. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 795.345733] env[62814]: DEBUG nova.network.neutron [req-69cc1953-0f59-4f5c-a4fa-cb10cb1f4f80 req-1b005d2f-661a-4686-bc93-6b8fdac25550 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updating instance_info_cache with network_info: [{"id": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "address": "fa:16:3e:aa:55:02", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b486fb-bb", "ovs_interfaceid": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.365692] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 795.365692] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52780c0e-d921-af2b-2a6e-3dc676767c8a" [ 795.365692] env[62814]: _type = "HttpNfcLease" [ 795.365692] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 795.366066] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 795.366066] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52780c0e-d921-af2b-2a6e-3dc676767c8a" [ 795.366066] env[62814]: _type = "HttpNfcLease" [ 795.366066] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 795.367329] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e3e93a-4ec3-4d4e-b33b-aaae0c213aa3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.375201] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521996c9-1286-c86a-7c65-e67b03571a2b/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 795.375456] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521996c9-1286-c86a-7c65-e67b03571a2b/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 795.440704] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c30d9d-822b-247e-3ad1-8081aa213f9b, 'name': SearchDatastore_Task, 'duration_secs': 0.010715} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.440704] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.440860] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 795.441074] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.441475] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 795.441475] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 795.441663] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7536deda-e7d3-43ba-b6ed-48fcbd3cec9a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.450443] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 795.450443] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 795.451151] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54cc038e-0609-4598-a86f-5ef29e11bd77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.456427] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 795.456427] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dba6f1-d10c-d9ce-e7db-a11a1337e787" [ 795.456427] env[62814]: _type = "Task" [ 795.456427] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.464932] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ed30fa07-37e9-4fa7-aa94-78f52e4ff70d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.473628] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dba6f1-d10c-d9ce-e7db-a11a1337e787, 'name': SearchDatastore_Task, 'duration_secs': 0.00921} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.473628] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cc96967-ae4b-4628-86a8-31e78878f4e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.478269] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 795.478269] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52423788-af72-b6d8-c152-57e2be66171d" [ 795.478269] env[62814]: _type = "Task" [ 795.478269] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.485952] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52423788-af72-b6d8-c152-57e2be66171d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.547102] env[62814]: DEBUG nova.compute.manager [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 795.547347] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 795.548788] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429ad87a-a67c-458f-ba5c-f6624e0f60bb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.556048] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 795.556265] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e94f96e9-a5ec-4f71-9ff7-b3e96235dda7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.563737] env[62814]: DEBUG oslo_vmware.api [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for the task: (returnval){ [ 795.563737] env[62814]: value = "task-4293643" [ 795.563737] env[62814]: _type = "Task" [ 795.563737] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.573903] env[62814]: DEBUG oslo_vmware.api [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293643, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.603550] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 795.620522] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "refresh_cache-cbc5985a-38e7-4e52-9fb0-264b5cec013a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.620684] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "refresh_cache-cbc5985a-38e7-4e52-9fb0-264b5cec013a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 795.620874] env[62814]: DEBUG nova.network.neutron [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 795.647924] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9449efb0-68e1-4a5d-8f0a-3fcf14d9c217 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "127fee64-fd56-4a23-bdd2-18c817898fd5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.459s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 795.686123] env[62814]: DEBUG nova.scheduler.client.report [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 68 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 795.686456] env[62814]: DEBUG nova.compute.provider_tree [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 68 to 69 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 795.686640] env[62814]: DEBUG nova.compute.provider_tree [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 795.801614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8a3df2f-497a-406d-b857-9187558f5111 tempest-ServerExternalEventsTest-781461774 tempest-ServerExternalEventsTest-781461774-project] Releasing lock "refresh_cache-b289800f-49ce-49e7-b6bc-a3b4ec84b434" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.849585] env[62814]: DEBUG oslo_concurrency.lockutils [req-69cc1953-0f59-4f5c-a4fa-cb10cb1f4f80 req-1b005d2f-661a-4686-bc93-6b8fdac25550 service nova] Releasing lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.995135] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52423788-af72-b6d8-c152-57e2be66171d, 'name': SearchDatastore_Task, 'duration_secs': 0.00953} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.996299] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.996299] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 3e4479a0-8edd-4b37-8cc9-2c91275b88ee/3e4479a0-8edd-4b37-8cc9-2c91275b88ee.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 795.996536] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-86dccdab-0128-4bd0-8954-c05abc100c84 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.007572] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 796.007572] env[62814]: value = "task-4293644" [ 796.007572] env[62814]: _type = "Task" [ 796.007572] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.018474] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293644, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.074719] env[62814]: DEBUG oslo_vmware.api [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293643, 'name': PowerOffVM_Task, 'duration_secs': 0.252253} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.075194] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 796.075704] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 796.075999] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-829c5ad7-4882-4a69-a07f-2651f8f27bbe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.142846] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 796.143120] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 796.145387] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Deleting the datastore file [datastore2] b289800f-49ce-49e7-b6bc-a3b4ec84b434 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 796.145387] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-017d81dd-0762-4d73-9530-c4ab0eba66d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.150112] env[62814]: DEBUG nova.compute.manager [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 796.154987] env[62814]: DEBUG oslo_vmware.api [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for the task: (returnval){ [ 796.154987] env[62814]: value = "task-4293646" [ 796.154987] env[62814]: _type = "Task" [ 796.154987] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.164263] env[62814]: DEBUG oslo_vmware.api [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293646, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.186101] env[62814]: DEBUG nova.network.neutron [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.195924] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.159s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.196546] env[62814]: DEBUG nova.compute.manager [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 796.201533] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.196s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.203211] env[62814]: INFO nova.compute.claims [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.522761] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293644, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.529157] env[62814]: DEBUG nova.network.neutron [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Updating instance_info_cache with network_info: [{"id": "1b565819-b3b3-4e33-b334-07d6fd8b9d62", "address": "fa:16:3e:9c:68:08", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b565819-b3", "ovs_interfaceid": "1b565819-b3b3-4e33-b334-07d6fd8b9d62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.662171] env[62814]: DEBUG nova.compute.manager [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Received event network-vif-plugged-1b565819-b3b3-4e33-b334-07d6fd8b9d62 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 796.663068] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] Acquiring lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.663068] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] Lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.663068] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] Lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.663251] env[62814]: DEBUG nova.compute.manager [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] No waiting events found dispatching network-vif-plugged-1b565819-b3b3-4e33-b334-07d6fd8b9d62 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 796.663251] env[62814]: WARNING nova.compute.manager [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Received unexpected event network-vif-plugged-1b565819-b3b3-4e33-b334-07d6fd8b9d62 for instance with vm_state building and task_state spawning. [ 796.663414] env[62814]: DEBUG nova.compute.manager [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Received event network-changed-1b565819-b3b3-4e33-b334-07d6fd8b9d62 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 796.663583] env[62814]: DEBUG nova.compute.manager [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Refreshing instance network info cache due to event network-changed-1b565819-b3b3-4e33-b334-07d6fd8b9d62. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 796.664405] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] Acquiring lock "refresh_cache-cbc5985a-38e7-4e52-9fb0-264b5cec013a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.682034] env[62814]: DEBUG oslo_vmware.api [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Task: {'id': task-4293646, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.469791} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.682034] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 796.682245] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 796.682364] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 796.682655] env[62814]: INFO nova.compute.manager [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Took 1.14 seconds to destroy the instance on the hypervisor. [ 796.682964] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 796.683216] env[62814]: DEBUG nova.compute.manager [-] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 796.683320] env[62814]: DEBUG nova.network.neutron [-] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 796.693725] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.714063] env[62814]: DEBUG nova.compute.utils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 796.716204] env[62814]: DEBUG nova.compute.manager [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 796.717951] env[62814]: DEBUG nova.network.neutron [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 797.021686] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293644, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541974} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.022299] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 3e4479a0-8edd-4b37-8cc9-2c91275b88ee/3e4479a0-8edd-4b37-8cc9-2c91275b88ee.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 797.022567] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 797.022939] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-675a832e-05c2-45f3-a493-25e9acfc8d2f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.033501] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 797.033501] env[62814]: value = "task-4293647" [ 797.033501] env[62814]: _type = "Task" [ 797.033501] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.035035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "refresh_cache-cbc5985a-38e7-4e52-9fb0-264b5cec013a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 797.035710] env[62814]: DEBUG nova.compute.manager [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Instance network_info: |[{"id": "1b565819-b3b3-4e33-b334-07d6fd8b9d62", "address": "fa:16:3e:9c:68:08", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b565819-b3", "ovs_interfaceid": "1b565819-b3b3-4e33-b334-07d6fd8b9d62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 797.036164] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] Acquired lock "refresh_cache-cbc5985a-38e7-4e52-9fb0-264b5cec013a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 797.036482] env[62814]: DEBUG nova.network.neutron [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Refreshing network info cache for port 1b565819-b3b3-4e33-b334-07d6fd8b9d62 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 797.038040] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9c:68:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1b565819-b3b3-4e33-b334-07d6fd8b9d62', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.049245] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 797.053217] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 797.056911] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fb54023c-4b16-4d5d-bde3-4fab549683a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.076528] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293647, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.081427] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.081427] env[62814]: value = "task-4293648" [ 797.081427] env[62814]: _type = "Task" [ 797.081427] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.090940] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293648, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.119609] env[62814]: DEBUG nova.policy [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '57ea9a44328f48608260b532bf3497a6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab4d42a6e2e54eaea18742dd5fb56c53', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 797.222781] env[62814]: DEBUG nova.compute.manager [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 797.270915] env[62814]: INFO nova.compute.manager [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Rebuilding instance [ 797.327202] env[62814]: DEBUG nova.compute.manager [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 797.328095] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf060fd3-3e8f-4ba5-afa1-6ab493e6243a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.545679] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293647, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.131681} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.548816] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.550073] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba83b26-a248-43f9-a943-da09f19f9193 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.579750] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 3e4479a0-8edd-4b37-8cc9-2c91275b88ee/3e4479a0-8edd-4b37-8cc9-2c91275b88ee.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.582791] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e187cc0-3dc7-4388-8bcc-6350eb6c11ad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.611221] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293648, 'name': CreateVM_Task, 'duration_secs': 0.52443} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.615203] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 797.617234] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 797.617234] env[62814]: value = "task-4293649" [ 797.617234] env[62814]: _type = "Task" [ 797.617234] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.617234] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.617234] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 797.617234] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 797.618973] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11debfb3-963b-4346-97d2-0a49b38f9af7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.628278] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 797.628278] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b5bf54-8f27-bed8-1ca9-220ee21a4189" [ 797.628278] env[62814]: _type = "Task" [ 797.628278] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.639899] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b5bf54-8f27-bed8-1ca9-220ee21a4189, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.765930] env[62814]: DEBUG nova.compute.manager [req-7289cbaa-3a17-4d16-89ab-41137c88ac84 req-17f46323-d509-4792-8946-dac765ef0f63 service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Received event network-vif-deleted-19fc1ad9-b48d-45ba-9f90-85b3a5910021 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 797.765930] env[62814]: INFO nova.compute.manager [req-7289cbaa-3a17-4d16-89ab-41137c88ac84 req-17f46323-d509-4792-8946-dac765ef0f63 service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Neutron deleted interface 19fc1ad9-b48d-45ba-9f90-85b3a5910021; detaching it from the instance and deleting it from the info cache [ 797.766226] env[62814]: DEBUG nova.network.neutron [req-7289cbaa-3a17-4d16-89ab-41137c88ac84 req-17f46323-d509-4792-8946-dac765ef0f63 service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.836612] env[62814]: DEBUG nova.network.neutron [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Updated VIF entry in instance network info cache for port 1b565819-b3b3-4e33-b334-07d6fd8b9d62. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 797.836967] env[62814]: DEBUG nova.network.neutron [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Updating instance_info_cache with network_info: [{"id": "1b565819-b3b3-4e33-b334-07d6fd8b9d62", "address": "fa:16:3e:9c:68:08", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1b565819-b3", "ovs_interfaceid": "1b565819-b3b3-4e33-b334-07d6fd8b9d62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.856571] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56be6320-7ee7-4ee2-bd7e-ffe5c86cd741 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.865758] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab60e2b-e0e2-4701-90a2-d5124433dbfc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.899780] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f8c5c8-9bea-4f50-8099-7bce590ca03e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.907487] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7835cfb6-f19e-4e93-ad15-087e43c9cd58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.922442] env[62814]: DEBUG nova.compute.provider_tree [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 797.937709] env[62814]: DEBUG nova.network.neutron [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Successfully created port: f5994178-e05c-464f-afea-a4f8a606efe3 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 798.138735] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293649, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.148789] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b5bf54-8f27-bed8-1ca9-220ee21a4189, 'name': SearchDatastore_Task, 'duration_secs': 0.016947} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.149020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 798.149270] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.149508] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.149733] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 798.149912] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 798.150195] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b89c3c4-194c-4e6a-ba78-e4a66c0a678a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.159707] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 798.159982] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 798.161079] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c572d41-32a0-4822-a785-900451c314db {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.167334] env[62814]: DEBUG nova.network.neutron [-] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.168423] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 798.168423] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ca30bf-1f7b-8df8-a886-9e273961d52c" [ 798.168423] env[62814]: _type = "Task" [ 798.168423] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.183451] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ca30bf-1f7b-8df8-a886-9e273961d52c, 'name': SearchDatastore_Task, 'duration_secs': 0.010853} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.184655] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ead0afe-b775-426f-b27b-5ea8896d0ce0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.190738] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 798.190738] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c6fcfc-61c4-92d4-b7fc-9f80e42d74ad" [ 798.190738] env[62814]: _type = "Task" [ 798.190738] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.199562] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c6fcfc-61c4-92d4-b7fc-9f80e42d74ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.241025] env[62814]: DEBUG nova.compute.manager [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 798.272684] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3762e70-6087-4d00-b031-3a9c932ff828 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.277558] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 798.277823] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.277984] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 798.278229] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.278309] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 798.278459] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 798.278665] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 798.278822] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 798.278989] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 798.282308] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 798.282505] env[62814]: DEBUG nova.virt.hardware [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 798.283383] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec87b28-d3cb-435f-b5b5-f7c5ac005936 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.290374] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc03d16-a56f-49a8-bc09-56389d98a4ac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.311017] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7acc3c-e288-4ea6-b935-d264dbf7f867 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.336608] env[62814]: DEBUG nova.compute.manager [req-7289cbaa-3a17-4d16-89ab-41137c88ac84 req-17f46323-d509-4792-8946-dac765ef0f63 service nova] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Detach interface failed, port_id=19fc1ad9-b48d-45ba-9f90-85b3a5910021, reason: Instance b289800f-49ce-49e7-b6bc-a3b4ec84b434 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 798.347215] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] Releasing lock "refresh_cache-cbc5985a-38e7-4e52-9fb0-264b5cec013a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 798.347527] env[62814]: DEBUG nova.compute.manager [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Received event network-changed-1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 798.347810] env[62814]: DEBUG nova.compute.manager [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Refreshing instance network info cache due to event network-changed-1865cb89-88d6-4e84-81ae-daece6adaa4d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 798.348084] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] Acquiring lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.348269] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] Acquired lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 798.348463] env[62814]: DEBUG nova.network.neutron [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Refreshing network info cache for port 1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 798.349914] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 798.350647] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77b89cff-f3e7-42f8-b296-7255c2cef907 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.358190] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 798.358190] env[62814]: value = "task-4293650" [ 798.358190] env[62814]: _type = "Task" [ 798.358190] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.372809] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293650, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.466843] env[62814]: DEBUG nova.scheduler.client.report [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 69 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 798.467248] env[62814]: DEBUG nova.compute.provider_tree [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 69 to 70 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 798.467536] env[62814]: DEBUG nova.compute.provider_tree [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 798.632835] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293649, 'name': ReconfigVM_Task, 'duration_secs': 0.955567} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.633306] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 3e4479a0-8edd-4b37-8cc9-2c91275b88ee/3e4479a0-8edd-4b37-8cc9-2c91275b88ee.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 798.633974] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-88c3ba81-7f5f-4bda-b98c-1f2701c6a1d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.641569] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 798.641569] env[62814]: value = "task-4293651" [ 798.641569] env[62814]: _type = "Task" [ 798.641569] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.654404] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293651, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.677131] env[62814]: INFO nova.compute.manager [-] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Took 1.99 seconds to deallocate network for instance. [ 798.701104] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c6fcfc-61c4-92d4-b7fc-9f80e42d74ad, 'name': SearchDatastore_Task, 'duration_secs': 0.010609} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.701307] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 798.703114] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] cbc5985a-38e7-4e52-9fb0-264b5cec013a/cbc5985a-38e7-4e52-9fb0-264b5cec013a.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 798.703114] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d75fc4fe-bbc8-4afb-b30d-c9ec0c81d8dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.707999] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 798.707999] env[62814]: value = "task-4293652" [ 798.707999] env[62814]: _type = "Task" [ 798.707999] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.720318] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293652, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.870773] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293650, 'name': PowerOffVM_Task, 'duration_secs': 0.176034} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.871209] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 798.871849] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 798.872732] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa381a39-af92-4ea6-bdb8-f5b8b636f384 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.879996] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 798.880286] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5cc3e040-096b-4de6-9c8a-913633f47ab6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.909271] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 798.909271] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 798.909271] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Deleting the datastore file [datastore2] 127fee64-fd56-4a23-bdd2-18c817898fd5 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 798.912153] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dc841cf8-a109-444d-8cb2-d78406a6b00c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.918968] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 798.918968] env[62814]: value = "task-4293654" [ 798.918968] env[62814]: _type = "Task" [ 798.918968] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.930542] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293654, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.976064] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.774s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.976637] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 798.979916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.587s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.979916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.979916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.193s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.979916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.984698] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.597s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.986692] env[62814]: INFO nova.compute.claims [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.017618] env[62814]: INFO nova.scheduler.client.report [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Deleted allocations for instance 961dedc6-5ddd-4620-bfa7-54cf0c520af3 [ 799.026320] env[62814]: INFO nova.scheduler.client.report [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Deleted allocations for instance 56f9e03c-1157-442f-8add-156627a6ec1e [ 799.156997] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293651, 'name': Rename_Task, 'duration_secs': 0.188582} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.156997] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 799.156997] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-106ae438-b128-4be8-8703-07f91a3024e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.164236] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 799.164236] env[62814]: value = "task-4293655" [ 799.164236] env[62814]: _type = "Task" [ 799.164236] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.175753] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293655, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.177170] env[62814]: DEBUG nova.network.neutron [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updated VIF entry in instance network info cache for port 1865cb89-88d6-4e84-81ae-daece6adaa4d. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 799.177570] env[62814]: DEBUG nova.network.neutron [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updating instance_info_cache with network_info: [{"id": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "address": "fa:16:3e:8e:bc:5f", "network": {"id": "ddc113f9-9fc4-4847-9181-82a3ee75ec5c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-787311776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6254d7e15b4c5ba267e8da03d1b82d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1865cb89-88", "ovs_interfaceid": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.184559] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.222356] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293652, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.430195] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293654, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.41447} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.430459] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 799.430712] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 799.430929] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 799.491189] env[62814]: DEBUG nova.compute.utils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 799.496202] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 799.496202] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 799.529964] env[62814]: DEBUG oslo_concurrency.lockutils [None req-147939e8-a0d6-4305-b16e-e20de0198ad5 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "961dedc6-5ddd-4620-bfa7-54cf0c520af3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.560s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.545662] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f80719a-a321-421e-8af2-7015a7b938a0 tempest-DeleteServersAdminTestJSON-627803745 tempest-DeleteServersAdminTestJSON-627803745-project-member] Lock "56f9e03c-1157-442f-8add-156627a6ec1e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.344s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.558724] env[62814]: DEBUG nova.policy [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24c9d7684648493c8e4beed961a51d3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eadee196b6264ba987043438ccdce5f8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 799.679912] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293655, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.680914] env[62814]: DEBUG oslo_concurrency.lockutils [req-8b185358-c46c-4068-a0b6-5bb5ae925e55 req-5011951b-920f-4713-89d9-b10f85d17cf1 service nova] Releasing lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 799.721773] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293652, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577333} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.723088] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] cbc5985a-38e7-4e52-9fb0-264b5cec013a/cbc5985a-38e7-4e52-9fb0-264b5cec013a.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 799.723499] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.724128] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bd165929-a1cf-497c-8ecf-136b236be0ac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.734230] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 799.734230] env[62814]: value = "task-4293656" [ 799.734230] env[62814]: _type = "Task" [ 799.734230] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.745026] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293656, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.873855] env[62814]: DEBUG nova.network.neutron [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Successfully updated port: f5994178-e05c-464f-afea-a4f8a606efe3 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.892254] env[62814]: DEBUG nova.compute.manager [req-da31b10c-b448-4b04-aa79-a9596b64535f req-c4c6d972-5a70-4cd9-9e00-5591af474b73 service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Received event network-vif-plugged-f5994178-e05c-464f-afea-a4f8a606efe3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 799.892254] env[62814]: DEBUG oslo_concurrency.lockutils [req-da31b10c-b448-4b04-aa79-a9596b64535f req-c4c6d972-5a70-4cd9-9e00-5591af474b73 service nova] Acquiring lock "1f40dc62-1a58-4cfb-8785-c37b68747f37-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 799.892891] env[62814]: DEBUG oslo_concurrency.lockutils [req-da31b10c-b448-4b04-aa79-a9596b64535f req-c4c6d972-5a70-4cd9-9e00-5591af474b73 service nova] Lock "1f40dc62-1a58-4cfb-8785-c37b68747f37-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 799.893258] env[62814]: DEBUG oslo_concurrency.lockutils [req-da31b10c-b448-4b04-aa79-a9596b64535f req-c4c6d972-5a70-4cd9-9e00-5591af474b73 service nova] Lock "1f40dc62-1a58-4cfb-8785-c37b68747f37-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.893640] env[62814]: DEBUG nova.compute.manager [req-da31b10c-b448-4b04-aa79-a9596b64535f req-c4c6d972-5a70-4cd9-9e00-5591af474b73 service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] No waiting events found dispatching network-vif-plugged-f5994178-e05c-464f-afea-a4f8a606efe3 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 799.893917] env[62814]: WARNING nova.compute.manager [req-da31b10c-b448-4b04-aa79-a9596b64535f req-c4c6d972-5a70-4cd9-9e00-5591af474b73 service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Received unexpected event network-vif-plugged-f5994178-e05c-464f-afea-a4f8a606efe3 for instance with vm_state building and task_state spawning. [ 800.000893] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 800.007024] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Successfully created port: 079f803d-e719-4778-8920-cecd6ea5510c {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 800.179629] env[62814]: DEBUG oslo_vmware.api [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293655, 'name': PowerOnVM_Task, 'duration_secs': 0.687482} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.183096] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 800.183096] env[62814]: INFO nova.compute.manager [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Took 8.92 seconds to spawn the instance on the hypervisor. [ 800.183096] env[62814]: DEBUG nova.compute.manager [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 800.183096] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6afcefb9-b1d2-4221-b9b6-f13e7897193e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.250924] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293656, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076012} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.253744] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 800.255534] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce1dd715-6f29-4b25-96b8-b13c6e0b7dee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.282768] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] cbc5985a-38e7-4e52-9fb0-264b5cec013a/cbc5985a-38e7-4e52-9fb0-264b5cec013a.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.285859] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b6f44d9-ebe6-4370-acc3-b675ba433318 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.306862] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 800.306862] env[62814]: value = "task-4293657" [ 800.306862] env[62814]: _type = "Task" [ 800.306862] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.317029] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293657, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.378176] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquiring lock "refresh_cache-1f40dc62-1a58-4cfb-8785-c37b68747f37" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.378176] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquired lock "refresh_cache-1f40dc62-1a58-4cfb-8785-c37b68747f37" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 800.378176] env[62814]: DEBUG nova.network.neutron [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 800.505907] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 800.505907] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.505907] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 800.505907] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.505907] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 800.505907] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 800.505907] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 800.505907] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 800.506540] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 800.506540] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 800.506540] env[62814]: DEBUG nova.virt.hardware [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 800.511559] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9442dc7-208d-467f-ae60-2687557a6bff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.522278] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f45fcf-7247-494b-8a2b-b0bf77bcac7f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.545266] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.552090] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 800.556082] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 800.556082] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c742fb6-4ebe-451e-8943-a55611eafda9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.581609] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.581609] env[62814]: value = "task-4293658" [ 800.581609] env[62814]: _type = "Task" [ 800.581609] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.592370] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293658, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.632168] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6752b2b6-d8f1-4429-aae6-a7a6777114b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.640672] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d6285e-912a-48e3-8c47-531ead6c3b19 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.689896] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-889143ad-e94a-4257-8a36-c29077e75852 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.702203] env[62814]: INFO nova.compute.manager [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Took 47.11 seconds to build instance. [ 800.703887] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37adffd5-7bb2-4742-90c1-6b7f27a79f2b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.720177] env[62814]: DEBUG nova.compute.provider_tree [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 800.818523] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293657, 'name': ReconfigVM_Task, 'duration_secs': 0.32587} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.818757] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Reconfigured VM instance instance-0000002c to attach disk [datastore2] cbc5985a-38e7-4e52-9fb0-264b5cec013a/cbc5985a-38e7-4e52-9fb0-264b5cec013a.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.820242] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d99e4f5-e7d8-4702-9811-3588b9388925 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.829151] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 800.829151] env[62814]: value = "task-4293659" [ 800.829151] env[62814]: _type = "Task" [ 800.829151] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.838598] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293659, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.961189] env[62814]: DEBUG nova.network.neutron [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.017641] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 801.071811] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 801.071811] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b0272f-4923-4f65-b1bb-9826e2a846a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.080748] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e220d50-3636-4105-9242-19094ab49979 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.102212] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293658, 'name': CreateVM_Task, 'duration_secs': 0.42981} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.102212] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 801.102685] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.102685] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 801.103184] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 801.103264] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dba2b19-3361-460e-b4cf-3edfb87117fb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.108116] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 801.108116] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525f0ccf-48c2-8488-7965-67e1988de567" [ 801.108116] env[62814]: _type = "Task" [ 801.108116] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.116513] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525f0ccf-48c2-8488-7965-67e1988de567, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.210829] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87204d3f-759d-4040-98ac-4cd341cfa755 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.949s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.222804] env[62814]: DEBUG nova.scheduler.client.report [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 801.257048] env[62814]: DEBUG nova.network.neutron [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Updating instance_info_cache with network_info: [{"id": "f5994178-e05c-464f-afea-a4f8a606efe3", "address": "fa:16:3e:ba:17:3e", "network": {"id": "12e384b1-79ae-4b53-8288-2b6be23e6640", "bridge": "br-int", "label": "tempest-ServersTestJSON-1101498603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab4d42a6e2e54eaea18742dd5fb56c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5994178-e0", "ovs_interfaceid": "f5994178-e05c-464f-afea-a4f8a606efe3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.348018] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293659, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.620626] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525f0ccf-48c2-8488-7965-67e1988de567, 'name': SearchDatastore_Task, 'duration_secs': 0.011064} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.621330] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 801.621330] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.621602] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.621765] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 801.621980] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.622327] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-94cd046b-f0c4-4eb1-9af8-4c891627156a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.641221] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquiring lock "448b7bfb-1a90-4c5b-9086-9320cedaf015" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 801.641221] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "448b7bfb-1a90-4c5b-9086-9320cedaf015" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 801.641221] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.641221] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 801.641668] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fc474f4-20a7-469d-a58d-be8d3cc17fdb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.648105] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 801.648105] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5260d96c-b2e3-d461-6272-34e3bb2da75a" [ 801.648105] env[62814]: _type = "Task" [ 801.648105] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.657967] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5260d96c-b2e3-d461-6272-34e3bb2da75a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.714045] env[62814]: DEBUG nova.compute.manager [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 801.733418] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.749s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.734077] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 801.737858] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.459s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 801.737858] env[62814]: DEBUG nova.objects.instance [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lazy-loading 'resources' on Instance uuid 961f313b-b43f-4531-8a4b-0a39421d6a34 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 801.760262] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Releasing lock "refresh_cache-1f40dc62-1a58-4cfb-8785-c37b68747f37" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 801.760585] env[62814]: DEBUG nova.compute.manager [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Instance network_info: |[{"id": "f5994178-e05c-464f-afea-a4f8a606efe3", "address": "fa:16:3e:ba:17:3e", "network": {"id": "12e384b1-79ae-4b53-8288-2b6be23e6640", "bridge": "br-int", "label": "tempest-ServersTestJSON-1101498603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab4d42a6e2e54eaea18742dd5fb56c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5994178-e0", "ovs_interfaceid": "f5994178-e05c-464f-afea-a4f8a606efe3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 801.762342] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:17:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '604056d6-6dd6-47fa-9eaa-6863a3a7c488', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5994178-e05c-464f-afea-a4f8a606efe3', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.776371] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Creating folder: Project (ab4d42a6e2e54eaea18742dd5fb56c53). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.778340] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4469383a-27f4-4f37-b449-1ad9ca2a2bf8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.784806] env[62814]: DEBUG nova.compute.manager [req-b5ba4563-a54e-4a39-86df-69bedbaff76b req-3d49bb5f-af01-4ebe-b450-cf4123062a25 service nova] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Received event network-vif-plugged-079f803d-e719-4778-8920-cecd6ea5510c {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 801.784806] env[62814]: DEBUG oslo_concurrency.lockutils [req-b5ba4563-a54e-4a39-86df-69bedbaff76b req-3d49bb5f-af01-4ebe-b450-cf4123062a25 service nova] Acquiring lock "2d0b24c5-3593-4ef2-a637-d3590242ad79-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 801.784900] env[62814]: DEBUG oslo_concurrency.lockutils [req-b5ba4563-a54e-4a39-86df-69bedbaff76b req-3d49bb5f-af01-4ebe-b450-cf4123062a25 service nova] Lock "2d0b24c5-3593-4ef2-a637-d3590242ad79-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 801.785022] env[62814]: DEBUG oslo_concurrency.lockutils [req-b5ba4563-a54e-4a39-86df-69bedbaff76b req-3d49bb5f-af01-4ebe-b450-cf4123062a25 service nova] Lock "2d0b24c5-3593-4ef2-a637-d3590242ad79-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.785196] env[62814]: DEBUG nova.compute.manager [req-b5ba4563-a54e-4a39-86df-69bedbaff76b req-3d49bb5f-af01-4ebe-b450-cf4123062a25 service nova] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] No waiting events found dispatching network-vif-plugged-079f803d-e719-4778-8920-cecd6ea5510c {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 801.785396] env[62814]: WARNING nova.compute.manager [req-b5ba4563-a54e-4a39-86df-69bedbaff76b req-3d49bb5f-af01-4ebe-b450-cf4123062a25 service nova] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Received unexpected event network-vif-plugged-079f803d-e719-4778-8920-cecd6ea5510c for instance with vm_state building and task_state spawning. [ 801.797946] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Created folder: Project (ab4d42a6e2e54eaea18742dd5fb56c53) in parent group-v845547. [ 801.798193] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Creating folder: Instances. Parent ref: group-v845676. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 801.798561] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4160481-c819-4976-bb96-9e52a92d7579 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.810258] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Created folder: Instances in parent group-v845676. [ 801.810568] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 801.810754] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 801.810979] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b6d13d4-1a09-443c-b066-072d56038918 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.831677] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.831677] env[62814]: value = "task-4293662" [ 801.831677] env[62814]: _type = "Task" [ 801.831677] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.842953] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293659, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.845817] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293662, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.890759] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Successfully updated port: 079f803d-e719-4778-8920-cecd6ea5510c {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.998662] env[62814]: DEBUG nova.compute.manager [req-571cd7bc-f801-4280-955d-4bd0e8043415 req-7a700242-7b3a-4f0e-a5ea-d71c98b564bb service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Received event network-changed-f5994178-e05c-464f-afea-a4f8a606efe3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 801.998798] env[62814]: DEBUG nova.compute.manager [req-571cd7bc-f801-4280-955d-4bd0e8043415 req-7a700242-7b3a-4f0e-a5ea-d71c98b564bb service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Refreshing instance network info cache due to event network-changed-f5994178-e05c-464f-afea-a4f8a606efe3. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 801.998959] env[62814]: DEBUG oslo_concurrency.lockutils [req-571cd7bc-f801-4280-955d-4bd0e8043415 req-7a700242-7b3a-4f0e-a5ea-d71c98b564bb service nova] Acquiring lock "refresh_cache-1f40dc62-1a58-4cfb-8785-c37b68747f37" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.002896] env[62814]: DEBUG oslo_concurrency.lockutils [req-571cd7bc-f801-4280-955d-4bd0e8043415 req-7a700242-7b3a-4f0e-a5ea-d71c98b564bb service nova] Acquired lock "refresh_cache-1f40dc62-1a58-4cfb-8785-c37b68747f37" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 802.002896] env[62814]: DEBUG nova.network.neutron [req-571cd7bc-f801-4280-955d-4bd0e8043415 req-7a700242-7b3a-4f0e-a5ea-d71c98b564bb service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Refreshing network info cache for port f5994178-e05c-464f-afea-a4f8a606efe3 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.159473] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5260d96c-b2e3-d461-6272-34e3bb2da75a, 'name': SearchDatastore_Task, 'duration_secs': 0.010875} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.160436] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81159121-bfe0-4bbd-8c7f-7e487e690bb3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.167084] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 802.167084] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52711c0e-90d8-4936-d3ce-776516cd89d8" [ 802.167084] env[62814]: _type = "Task" [ 802.167084] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.187628] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52711c0e-90d8-4936-d3ce-776516cd89d8, 'name': SearchDatastore_Task, 'duration_secs': 0.01264} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.190869] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 802.190869] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 127fee64-fd56-4a23-bdd2-18c817898fd5/127fee64-fd56-4a23-bdd2-18c817898fd5.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 802.190869] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b122e6d1-8e40-4a16-80d5-fca9a30aba54 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.195701] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 802.195701] env[62814]: value = "task-4293663" [ 802.195701] env[62814]: _type = "Task" [ 802.195701] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.205843] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293663, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.244321] env[62814]: DEBUG nova.compute.utils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 802.247158] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 802.248177] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 802.250688] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 802.315188] env[62814]: DEBUG nova.policy [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24c9d7684648493c8e4beed961a51d3f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eadee196b6264ba987043438ccdce5f8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 802.343261] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293659, 'name': Rename_Task, 'duration_secs': 1.179267} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.347000] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 802.347327] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293662, 'name': CreateVM_Task, 'duration_secs': 0.430917} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.350330] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-401915a7-b919-4f06-8515-5e9a4a81c6c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.352857] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 802.354563] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.354796] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 802.355170] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 802.355859] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b005d6dd-a873-459d-a518-ce0c510deda1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.360422] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 802.360422] env[62814]: value = "task-4293664" [ 802.360422] env[62814]: _type = "Task" [ 802.360422] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.361980] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for the task: (returnval){ [ 802.361980] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5282541f-031d-2a50-ef53-a6e011b40032" [ 802.361980] env[62814]: _type = "Task" [ 802.361980] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.378937] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293664, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.383177] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5282541f-031d-2a50-ef53-a6e011b40032, 'name': SearchDatastore_Task, 'duration_secs': 0.011278} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.383935] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 802.383935] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.384223] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.384278] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 802.385075] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.385075] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34d39ee3-3f04-42c6-93c8-c0cd8665bd88 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.394575] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "refresh_cache-2d0b24c5-3593-4ef2-a637-d3590242ad79" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.394722] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "refresh_cache-2d0b24c5-3593-4ef2-a637-d3590242ad79" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 802.394876] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 802.404833] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.405072] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 802.405862] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acacaadc-adc3-4437-9384-6708333a195a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.417894] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for the task: (returnval){ [ 802.417894] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52134fb4-18f2-4c4d-678a-178de0010689" [ 802.417894] env[62814]: _type = "Task" [ 802.417894] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.432192] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52134fb4-18f2-4c4d-678a-178de0010689, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.708862] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293663, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.753974] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 802.879428] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293664, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.903668] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Successfully created port: d79a3b82-ece8-4a3e-9549-734213a374c7 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 802.929192] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52134fb4-18f2-4c4d-678a-178de0010689, 'name': SearchDatastore_Task, 'duration_secs': 0.0879} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.934065] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e2cec6e-e620-4e2c-9275-aef07ae5ec44 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.940164] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for the task: (returnval){ [ 802.940164] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dd2029-8111-4717-5786-e680e756be5e" [ 802.940164] env[62814]: _type = "Task" [ 802.940164] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.954942] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dd2029-8111-4717-5786-e680e756be5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.984749] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.991034] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2cac241-6a89-43e4-a4ee-b58c64e1b97e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.999434] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d815a7c-36cd-4d96-b444-b047558e8c74 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.040031] env[62814]: DEBUG nova.network.neutron [req-571cd7bc-f801-4280-955d-4bd0e8043415 req-7a700242-7b3a-4f0e-a5ea-d71c98b564bb service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Updated VIF entry in instance network info cache for port f5994178-e05c-464f-afea-a4f8a606efe3. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 803.040031] env[62814]: DEBUG nova.network.neutron [req-571cd7bc-f801-4280-955d-4bd0e8043415 req-7a700242-7b3a-4f0e-a5ea-d71c98b564bb service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Updating instance_info_cache with network_info: [{"id": "f5994178-e05c-464f-afea-a4f8a606efe3", "address": "fa:16:3e:ba:17:3e", "network": {"id": "12e384b1-79ae-4b53-8288-2b6be23e6640", "bridge": "br-int", "label": "tempest-ServersTestJSON-1101498603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab4d42a6e2e54eaea18742dd5fb56c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5994178-e0", "ovs_interfaceid": "f5994178-e05c-464f-afea-a4f8a606efe3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.045078] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f41ebe-f874-4f1a-8826-c4c86e6ad652 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.052949] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26afa124-8b75-4642-b59d-4e104de72683 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.069161] env[62814]: DEBUG nova.compute.provider_tree [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 803.210416] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293663, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589488} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.210416] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 127fee64-fd56-4a23-bdd2-18c817898fd5/127fee64-fd56-4a23-bdd2-18c817898fd5.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 803.210416] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.210416] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4594223-00ca-4f10-8c1f-aec2303c49bb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.216577] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 803.216577] env[62814]: value = "task-4293665" [ 803.216577] env[62814]: _type = "Task" [ 803.216577] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.224708] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.365964] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Updating instance_info_cache with network_info: [{"id": "079f803d-e719-4778-8920-cecd6ea5510c", "address": "fa:16:3e:89:33:b0", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap079f803d-e7", "ovs_interfaceid": "079f803d-e719-4778-8920-cecd6ea5510c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.382757] env[62814]: DEBUG oslo_vmware.api [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293664, 'name': PowerOnVM_Task, 'duration_secs': 0.849647} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.383053] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 803.383296] env[62814]: INFO nova.compute.manager [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Took 9.33 seconds to spawn the instance on the hypervisor. [ 803.383476] env[62814]: DEBUG nova.compute.manager [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 803.384340] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba3b75f-a4ad-422d-a166-6d941a291be2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.454634] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dd2029-8111-4717-5786-e680e756be5e, 'name': SearchDatastore_Task, 'duration_secs': 0.010083} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.458234] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.458837] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1f40dc62-1a58-4cfb-8785-c37b68747f37/1f40dc62-1a58-4cfb-8785-c37b68747f37.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 803.459916] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1eae585e-628f-4125-8ca5-251cb3c7bf2d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.470885] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for the task: (returnval){ [ 803.470885] env[62814]: value = "task-4293666" [ 803.470885] env[62814]: _type = "Task" [ 803.470885] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.480285] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293666, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.548720] env[62814]: DEBUG oslo_concurrency.lockutils [req-571cd7bc-f801-4280-955d-4bd0e8043415 req-7a700242-7b3a-4f0e-a5ea-d71c98b564bb service nova] Releasing lock "refresh_cache-1f40dc62-1a58-4cfb-8785-c37b68747f37" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.600577] env[62814]: ERROR nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [req-638f9820-717f-4c22-9eeb-a4ee99c7c9d4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-638f9820-717f-4c22-9eeb-a4ee99c7c9d4"}]} [ 803.625498] env[62814]: DEBUG nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 803.648178] env[62814]: DEBUG nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 803.648423] env[62814]: DEBUG nova.compute.provider_tree [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 803.665641] env[62814]: DEBUG nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 803.697542] env[62814]: DEBUG nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 803.728308] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077035} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.730017] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.732056] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a116fa22-8328-48ed-9ab5-4ce07242ff65 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.756126] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 127fee64-fd56-4a23-bdd2-18c817898fd5/127fee64-fd56-4a23-bdd2-18c817898fd5.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.759482] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ece463b-50ab-403c-a276-8add24ec1e56 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.775865] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 803.782858] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 803.782858] env[62814]: value = "task-4293667" [ 803.782858] env[62814]: _type = "Task" [ 803.782858] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.793925] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293667, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.806123] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 803.806580] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 803.806970] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 803.807050] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 803.807504] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 803.807572] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 803.807845] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 803.808037] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 803.808250] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 803.808754] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 803.808754] env[62814]: DEBUG nova.virt.hardware [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 803.809872] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40319292-3959-4bc2-b0a8-f3a479985cab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.818224] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71152457-e706-4117-8960-accbd45b1226 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.878076] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "refresh_cache-2d0b24c5-3593-4ef2-a637-d3590242ad79" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 803.878484] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Instance network_info: |[{"id": "079f803d-e719-4778-8920-cecd6ea5510c", "address": "fa:16:3e:89:33:b0", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap079f803d-e7", "ovs_interfaceid": "079f803d-e719-4778-8920-cecd6ea5510c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 803.878924] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:33:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '079f803d-e719-4778-8920-cecd6ea5510c', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 803.888709] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 803.890281] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 803.890708] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3fd9e01c-1c51-4acc-ac95-1c5bef1c17be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.920575] env[62814]: INFO nova.compute.manager [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Took 42.88 seconds to build instance. [ 803.924474] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 803.924474] env[62814]: value = "task-4293668" [ 803.924474] env[62814]: _type = "Task" [ 803.924474] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.934736] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293668, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.981869] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293666, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.109328] env[62814]: DEBUG nova.compute.manager [req-f904c1a3-9114-49c5-b266-6964172e2326 req-3f7c51e5-7e61-47ef-9a50-b1cf0a9a7d66 service nova] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Received event network-changed-079f803d-e719-4778-8920-cecd6ea5510c {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 804.109617] env[62814]: DEBUG nova.compute.manager [req-f904c1a3-9114-49c5-b266-6964172e2326 req-3f7c51e5-7e61-47ef-9a50-b1cf0a9a7d66 service nova] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Refreshing instance network info cache due to event network-changed-079f803d-e719-4778-8920-cecd6ea5510c. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 804.110552] env[62814]: DEBUG oslo_concurrency.lockutils [req-f904c1a3-9114-49c5-b266-6964172e2326 req-3f7c51e5-7e61-47ef-9a50-b1cf0a9a7d66 service nova] Acquiring lock "refresh_cache-2d0b24c5-3593-4ef2-a637-d3590242ad79" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.110790] env[62814]: DEBUG oslo_concurrency.lockutils [req-f904c1a3-9114-49c5-b266-6964172e2326 req-3f7c51e5-7e61-47ef-9a50-b1cf0a9a7d66 service nova] Acquired lock "refresh_cache-2d0b24c5-3593-4ef2-a637-d3590242ad79" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 804.111053] env[62814]: DEBUG nova.network.neutron [req-f904c1a3-9114-49c5-b266-6964172e2326 req-3f7c51e5-7e61-47ef-9a50-b1cf0a9a7d66 service nova] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Refreshing network info cache for port 079f803d-e719-4778-8920-cecd6ea5510c {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.298493] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293667, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.385019] env[62814]: DEBUG nova.compute.manager [req-05a1b021-cdc5-4fd5-85f0-934ab17f8f2f req-fc54e556-6f8d-4fa8-9af0-7cc3da909e88 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Received event network-changed-95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 804.385019] env[62814]: DEBUG nova.compute.manager [req-05a1b021-cdc5-4fd5-85f0-934ab17f8f2f req-fc54e556-6f8d-4fa8-9af0-7cc3da909e88 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Refreshing instance network info cache due to event network-changed-95b486fb-bbe0-4666-a926-01bcfa37eaad. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 804.385019] env[62814]: DEBUG oslo_concurrency.lockutils [req-05a1b021-cdc5-4fd5-85f0-934ab17f8f2f req-fc54e556-6f8d-4fa8-9af0-7cc3da909e88 service nova] Acquiring lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.385408] env[62814]: DEBUG oslo_concurrency.lockutils [req-05a1b021-cdc5-4fd5-85f0-934ab17f8f2f req-fc54e556-6f8d-4fa8-9af0-7cc3da909e88 service nova] Acquired lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 804.385408] env[62814]: DEBUG nova.network.neutron [req-05a1b021-cdc5-4fd5-85f0-934ab17f8f2f req-fc54e556-6f8d-4fa8-9af0-7cc3da909e88 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Refreshing network info cache for port 95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.406183] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d37c4fc4-be9b-4456-935e-51602ad97f0b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.415038] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3464785-c0ff-46b8-854d-162a5bc70aea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.451295] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed48905f-2bae-415e-8650-b1bc26940c49 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.993s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 804.457111] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70115626-7548-4a12-bea2-6c2c39c0ce94 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.466620] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293668, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.470205] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-412ba1cd-d869-4a07-be65-02c76ab1f814 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.489950] env[62814]: DEBUG nova.compute.provider_tree [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 804.497558] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293666, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.836698} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.498318] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1f40dc62-1a58-4cfb-8785-c37b68747f37/1f40dc62-1a58-4cfb-8785-c37b68747f37.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 804.498318] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.498483] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8607bcc-a42a-485f-8b9c-6db9f8e3bdeb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.506905] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for the task: (returnval){ [ 804.506905] env[62814]: value = "task-4293669" [ 804.506905] env[62814]: _type = "Task" [ 804.506905] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.516930] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293669, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.774974] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 804.775249] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 804.802755] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293667, 'name': ReconfigVM_Task, 'duration_secs': 0.538639} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.803053] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 127fee64-fd56-4a23-bdd2-18c817898fd5/127fee64-fd56-4a23-bdd2-18c817898fd5.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.803802] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-abd5ddd5-0233-4d22-8ed2-3d86a4017a61 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.811889] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 804.811889] env[62814]: value = "task-4293670" [ 804.811889] env[62814]: _type = "Task" [ 804.811889] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.820763] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293670, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.881895] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Successfully updated port: d79a3b82-ece8-4a3e-9549-734213a374c7 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 804.959681] env[62814]: DEBUG nova.compute.manager [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 804.970129] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293668, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.976841] env[62814]: DEBUG nova.network.neutron [req-f904c1a3-9114-49c5-b266-6964172e2326 req-3f7c51e5-7e61-47ef-9a50-b1cf0a9a7d66 service nova] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Updated VIF entry in instance network info cache for port 079f803d-e719-4778-8920-cecd6ea5510c. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 804.976841] env[62814]: DEBUG nova.network.neutron [req-f904c1a3-9114-49c5-b266-6964172e2326 req-3f7c51e5-7e61-47ef-9a50-b1cf0a9a7d66 service nova] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Updating instance_info_cache with network_info: [{"id": "079f803d-e719-4778-8920-cecd6ea5510c", "address": "fa:16:3e:89:33:b0", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap079f803d-e7", "ovs_interfaceid": "079f803d-e719-4778-8920-cecd6ea5510c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.018427] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293669, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07705} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.019853] env[62814]: ERROR nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] [req-43a6a77b-de8d-47f6-b3dd-27f05d2badcc] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-43a6a77b-de8d-47f6-b3dd-27f05d2badcc"}]} [ 805.020955] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 805.025325] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4809d112-fe3e-4ac3-8a16-7a13cf8aa40c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.053024] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 1f40dc62-1a58-4cfb-8785-c37b68747f37/1f40dc62-1a58-4cfb-8785-c37b68747f37.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.054689] env[62814]: DEBUG nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 805.057311] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1bde615-53a6-446e-aac2-ca6a52c4a100 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.077873] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "a473cc7e-0f86-4a18-9789-0aabfff430ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 805.078063] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "a473cc7e-0f86-4a18-9789-0aabfff430ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 805.082971] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for the task: (returnval){ [ 805.082971] env[62814]: value = "task-4293671" [ 805.082971] env[62814]: _type = "Task" [ 805.082971] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.094979] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293671, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.096467] env[62814]: DEBUG nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 805.096467] env[62814]: DEBUG nova.compute.provider_tree [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 805.108299] env[62814]: DEBUG nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 805.136099] env[62814]: DEBUG nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 805.321483] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293670, 'name': Rename_Task, 'duration_secs': 0.155529} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.324283] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 805.324790] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd16cd5a-c054-4dc0-ad79-01031e731e19 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.332837] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 805.332837] env[62814]: value = "task-4293672" [ 805.332837] env[62814]: _type = "Task" [ 805.332837] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.345799] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293672, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.354234] env[62814]: DEBUG nova.network.neutron [req-05a1b021-cdc5-4fd5-85f0-934ab17f8f2f req-fc54e556-6f8d-4fa8-9af0-7cc3da909e88 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updated VIF entry in instance network info cache for port 95b486fb-bbe0-4666-a926-01bcfa37eaad. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 805.354234] env[62814]: DEBUG nova.network.neutron [req-05a1b021-cdc5-4fd5-85f0-934ab17f8f2f req-fc54e556-6f8d-4fa8-9af0-7cc3da909e88 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updating instance_info_cache with network_info: [{"id": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "address": "fa:16:3e:aa:55:02", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b486fb-bb", "ovs_interfaceid": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.387417] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "refresh_cache-c18acec7-cf95-4cdf-aa49-32419d364534" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.387720] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "refresh_cache-c18acec7-cf95-4cdf-aa49-32419d364534" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 805.387720] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 805.470412] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293668, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.472447] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521996c9-1286-c86a-7c65-e67b03571a2b/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 805.477365] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e224340-3795-44c3-b645-c2aefadad049 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.480408] env[62814]: DEBUG oslo_concurrency.lockutils [req-f904c1a3-9114-49c5-b266-6964172e2326 req-3f7c51e5-7e61-47ef-9a50-b1cf0a9a7d66 service nova] Releasing lock "refresh_cache-2d0b24c5-3593-4ef2-a637-d3590242ad79" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 805.484435] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521996c9-1286-c86a-7c65-e67b03571a2b/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 805.484617] env[62814]: ERROR oslo_vmware.rw_handles [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521996c9-1286-c86a-7c65-e67b03571a2b/disk-0.vmdk due to incomplete transfer. [ 805.484871] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-262c908c-0c3d-4a01-bdd1-9a22136d2606 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.491718] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 805.495878] env[62814]: DEBUG oslo_vmware.rw_handles [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521996c9-1286-c86a-7c65-e67b03571a2b/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 805.496486] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Uploaded image 1a74e7e6-0277-48fd-9c78-c75a4976ef3e to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 805.498478] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 805.498758] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8656bdd3-014a-47f2-bf93-ecd1bf9795d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.509849] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 805.509849] env[62814]: value = "task-4293673" [ 805.509849] env[62814]: _type = "Task" [ 805.509849] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.519885] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293673, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.594860] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293671, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.705818] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73c54c52-b8ec-45c3-bd6e-3c887d38aac1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.713938] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65de5e2-af50-44d8-8da3-3fe85559174a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.744458] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5264b587-9992-4a7e-9b59-456dd2e12f47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.752327] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5cf8787-9c9c-497f-972c-c20ccd0f85c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.766972] env[62814]: DEBUG nova.compute.provider_tree [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 805.843971] env[62814]: DEBUG oslo_vmware.api [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293672, 'name': PowerOnVM_Task, 'duration_secs': 0.47526} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.844289] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 805.844856] env[62814]: DEBUG nova.compute.manager [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 805.845625] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205097bd-8e26-4333-a78a-842be6d3a440 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.856779] env[62814]: DEBUG oslo_concurrency.lockutils [req-05a1b021-cdc5-4fd5-85f0-934ab17f8f2f req-fc54e556-6f8d-4fa8-9af0-7cc3da909e88 service nova] Releasing lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 805.942575] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.969041] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293668, 'name': CreateVM_Task, 'duration_secs': 1.735556} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.971764] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 805.972598] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.973290] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 805.973290] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 805.973934] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98189054-0605-4ef4-bd8a-d25bce48694b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.979522] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 805.979522] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d18702-e410-fb42-e9f2-0de8ab896dbe" [ 805.979522] env[62814]: _type = "Task" [ 805.979522] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.988752] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d18702-e410-fb42-e9f2-0de8ab896dbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.022279] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293673, 'name': Destroy_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.095982] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293671, 'name': ReconfigVM_Task, 'duration_secs': 0.855777} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.096299] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 1f40dc62-1a58-4cfb-8785-c37b68747f37/1f40dc62-1a58-4cfb-8785-c37b68747f37.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 806.096924] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-520c097d-01b3-4e53-a610-526ef7accc8c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.107331] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for the task: (returnval){ [ 806.107331] env[62814]: value = "task-4293674" [ 806.107331] env[62814]: _type = "Task" [ 806.107331] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.114022] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293674, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.184675] env[62814]: DEBUG nova.network.neutron [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Updating instance_info_cache with network_info: [{"id": "d79a3b82-ece8-4a3e-9549-734213a374c7", "address": "fa:16:3e:a1:34:62", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd79a3b82-ec", "ovs_interfaceid": "d79a3b82-ece8-4a3e-9549-734213a374c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.206969] env[62814]: DEBUG nova.compute.manager [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Received event network-vif-plugged-d79a3b82-ece8-4a3e-9549-734213a374c7 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 806.207231] env[62814]: DEBUG oslo_concurrency.lockutils [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] Acquiring lock "c18acec7-cf95-4cdf-aa49-32419d364534-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.207456] env[62814]: DEBUG oslo_concurrency.lockutils [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] Lock "c18acec7-cf95-4cdf-aa49-32419d364534-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.207597] env[62814]: DEBUG oslo_concurrency.lockutils [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] Lock "c18acec7-cf95-4cdf-aa49-32419d364534-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.207822] env[62814]: DEBUG nova.compute.manager [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] No waiting events found dispatching network-vif-plugged-d79a3b82-ece8-4a3e-9549-734213a374c7 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 806.207913] env[62814]: WARNING nova.compute.manager [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Received unexpected event network-vif-plugged-d79a3b82-ece8-4a3e-9549-734213a374c7 for instance with vm_state building and task_state spawning. [ 806.208783] env[62814]: DEBUG nova.compute.manager [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Received event network-changed-d79a3b82-ece8-4a3e-9549-734213a374c7 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 806.209051] env[62814]: DEBUG nova.compute.manager [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Refreshing instance network info cache due to event network-changed-d79a3b82-ece8-4a3e-9549-734213a374c7. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 806.209204] env[62814]: DEBUG oslo_concurrency.lockutils [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] Acquiring lock "refresh_cache-c18acec7-cf95-4cdf-aa49-32419d364534" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.307214] env[62814]: DEBUG nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 73 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 806.307214] env[62814]: DEBUG nova.compute.provider_tree [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 73 to 74 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 806.307214] env[62814]: DEBUG nova.compute.provider_tree [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 806.365425] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.490192] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d18702-e410-fb42-e9f2-0de8ab896dbe, 'name': SearchDatastore_Task, 'duration_secs': 0.012444} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.490648] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 806.490648] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.490808] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.490989] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 806.491248] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.491479] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-028a1dac-60d3-47c9-aeaa-cefacadd4d11 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.502540] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.502766] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 806.503762] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36b2dca6-7975-4353-b7e5-3c3f53e1c021 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.509429] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 806.509429] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f00c68-d80f-f3ce-e788-6eec262bfbcc" [ 806.509429] env[62814]: _type = "Task" [ 806.509429] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.520109] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f00c68-d80f-f3ce-e788-6eec262bfbcc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.523441] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293673, 'name': Destroy_Task, 'duration_secs': 0.524834} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.523970] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Destroyed the VM [ 806.524228] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 806.524589] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-427b0747-2196-4c43-9d9c-82402472a49b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.529823] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 806.529823] env[62814]: value = "task-4293675" [ 806.529823] env[62814]: _type = "Task" [ 806.529823] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.540770] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293675, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.614519] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293674, 'name': Rename_Task, 'duration_secs': 0.471719} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.614997] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 806.615269] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2e44a22-edc7-4090-b4c2-55ee9b4f573b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.620935] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for the task: (returnval){ [ 806.620935] env[62814]: value = "task-4293676" [ 806.620935] env[62814]: _type = "Task" [ 806.620935] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.631125] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293676, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.687954] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "refresh_cache-c18acec7-cf95-4cdf-aa49-32419d364534" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 806.688329] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Instance network_info: |[{"id": "d79a3b82-ece8-4a3e-9549-734213a374c7", "address": "fa:16:3e:a1:34:62", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd79a3b82-ec", "ovs_interfaceid": "d79a3b82-ece8-4a3e-9549-734213a374c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 806.688632] env[62814]: DEBUG oslo_concurrency.lockutils [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] Acquired lock "refresh_cache-c18acec7-cf95-4cdf-aa49-32419d364534" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 806.688809] env[62814]: DEBUG nova.network.neutron [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Refreshing network info cache for port d79a3b82-ece8-4a3e-9549-734213a374c7 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 806.690164] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:34:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cca1f087-01e1-49ca-831b-5c51478a5d60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd79a3b82-ece8-4a3e-9549-734213a374c7', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 806.698904] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 806.699544] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 806.700240] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c1f95d3-e8ab-4d75-8e37-bc4fb2e48665 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.723799] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 806.723799] env[62814]: value = "task-4293677" [ 806.723799] env[62814]: _type = "Task" [ 806.723799] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.737022] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293677, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.811975] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 5.074s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.814956] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.402s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.815230] env[62814]: DEBUG nova.objects.instance [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lazy-loading 'resources' on Instance uuid 64a475e4-6713-408b-a63a-a43b5fed5ec8 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 806.882990] env[62814]: INFO nova.scheduler.client.report [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Deleted allocations for instance 961f313b-b43f-4531-8a4b-0a39421d6a34 [ 807.027511] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f00c68-d80f-f3ce-e788-6eec262bfbcc, 'name': SearchDatastore_Task, 'duration_secs': 0.020781} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.028319] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88b47c4e-e567-44be-9bba-d9eb9100f1b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.038285] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 807.038285] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a12ca3-59dc-fb9c-bd54-218cc9e24cd9" [ 807.038285] env[62814]: _type = "Task" [ 807.038285] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.046689] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293675, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.052693] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a12ca3-59dc-fb9c-bd54-218cc9e24cd9, 'name': SearchDatastore_Task} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.053054] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 807.053333] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 2d0b24c5-3593-4ef2-a637-d3590242ad79/2d0b24c5-3593-4ef2-a637-d3590242ad79.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 807.053662] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-73747dc7-6df4-44fd-a0ec-a4457c8ce5da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.060238] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 807.060238] env[62814]: value = "task-4293678" [ 807.060238] env[62814]: _type = "Task" [ 807.060238] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.069324] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293678, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.135123] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293676, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.235553] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293677, 'name': CreateVM_Task, 'duration_secs': 0.361697} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.237167] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 807.237977] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.238215] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 807.238984] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 807.238984] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c43889e-6d0a-4370-ab1c-f6d993fdec66 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.246458] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 807.246458] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525185a0-2abc-ced1-94b2-50b6ad1fe312" [ 807.246458] env[62814]: _type = "Task" [ 807.246458] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.260345] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525185a0-2abc-ced1-94b2-50b6ad1fe312, 'name': SearchDatastore_Task, 'duration_secs': 0.010635} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.261221] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 807.261221] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.261221] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.261487] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 807.261688] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.263078] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-84cb65c8-0cc7-40a4-a49a-b8712a172e4a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.271280] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.271559] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 807.272340] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49730b55-2fa4-485b-a7ec-a77bb3fd261a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.278226] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 807.278226] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b77f6-855b-1262-d91b-be1b6bc0166f" [ 807.278226] env[62814]: _type = "Task" [ 807.278226] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.287075] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b77f6-855b-1262-d91b-be1b6bc0166f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.392852] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b92f45b-862d-456e-9d59-ea6d827d37e1 tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "961f313b-b43f-4531-8a4b-0a39421d6a34" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.879s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 807.527357] env[62814]: DEBUG nova.network.neutron [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Updated VIF entry in instance network info cache for port d79a3b82-ece8-4a3e-9549-734213a374c7. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 807.527845] env[62814]: DEBUG nova.network.neutron [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Updating instance_info_cache with network_info: [{"id": "d79a3b82-ece8-4a3e-9549-734213a374c7", "address": "fa:16:3e:a1:34:62", "network": {"id": "ede01690-e2ec-41a2-a0e9-3abcf604935f", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-445344456-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eadee196b6264ba987043438ccdce5f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cca1f087-01e1-49ca-831b-5c51478a5d60", "external-id": "nsx-vlan-transportzone-439", "segmentation_id": 439, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd79a3b82-ec", "ovs_interfaceid": "d79a3b82-ece8-4a3e-9549-734213a374c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.546834] env[62814]: DEBUG oslo_vmware.api [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293675, 'name': RemoveSnapshot_Task, 'duration_secs': 0.949657} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.552954] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 807.552954] env[62814]: INFO nova.compute.manager [None req-dc97194c-741e-4b85-a6b9-65cc7998b1a6 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Took 15.86 seconds to snapshot the instance on the hypervisor. [ 807.578739] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293678, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.586997] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "127fee64-fd56-4a23-bdd2-18c817898fd5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 807.586997] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "127fee64-fd56-4a23-bdd2-18c817898fd5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 807.586997] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "127fee64-fd56-4a23-bdd2-18c817898fd5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 807.587217] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "127fee64-fd56-4a23-bdd2-18c817898fd5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 807.591107] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "127fee64-fd56-4a23-bdd2-18c817898fd5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 807.595019] env[62814]: INFO nova.compute.manager [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Terminating instance [ 807.632813] env[62814]: DEBUG oslo_vmware.api [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293676, 'name': PowerOnVM_Task, 'duration_secs': 0.88052} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.632997] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 807.633163] env[62814]: INFO nova.compute.manager [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Took 9.39 seconds to spawn the instance on the hypervisor. [ 807.633374] env[62814]: DEBUG nova.compute.manager [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 807.634228] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c110780-8164-4aae-86b2-cd9694e9ea19 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.790179] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b77f6-855b-1262-d91b-be1b6bc0166f, 'name': SearchDatastore_Task, 'duration_secs': 0.00929} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.793264] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c44dfcbf-355a-49b2-a361-5c06731273bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.801894] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 807.801894] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5234eefc-37f9-88da-5789-c225f0264d9a" [ 807.801894] env[62814]: _type = "Task" [ 807.801894] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.811771] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5234eefc-37f9-88da-5789-c225f0264d9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.970685] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e29ae60-db27-45bf-89e6-c9f7243ea6f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.979521] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74ba21a-db10-4da4-bcf5-c9f42eb53d66 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.014721] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55dd0d83-63a0-403f-b407-d5b21ea64bc4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.025778] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6461c597-202c-4c38-8214-9435c36060ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.030891] env[62814]: DEBUG oslo_concurrency.lockutils [req-e026f76e-fabe-41ae-b94d-67c79130518e req-7ec0fe28-3434-4e3f-b849-39bcce603c2a service nova] Releasing lock "refresh_cache-c18acec7-cf95-4cdf-aa49-32419d364534" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 808.042839] env[62814]: DEBUG nova.compute.provider_tree [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.076471] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293678, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.635557} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.076825] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 2d0b24c5-3593-4ef2-a637-d3590242ad79/2d0b24c5-3593-4ef2-a637-d3590242ad79.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 808.077073] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 808.077324] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ca38c79-4165-4871-9253-11f799e423f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.087343] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 808.087343] env[62814]: value = "task-4293679" [ 808.087343] env[62814]: _type = "Task" [ 808.087343] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.097742] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293679, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.098534] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "refresh_cache-127fee64-fd56-4a23-bdd2-18c817898fd5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.098701] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquired lock "refresh_cache-127fee64-fd56-4a23-bdd2-18c817898fd5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 808.098874] env[62814]: DEBUG nova.network.neutron [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 808.163247] env[62814]: INFO nova.compute.manager [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Took 43.78 seconds to build instance. [ 808.315926] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5234eefc-37f9-88da-5789-c225f0264d9a, 'name': SearchDatastore_Task, 'duration_secs': 0.025623} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.316304] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 808.316619] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c18acec7-cf95-4cdf-aa49-32419d364534/c18acec7-cf95-4cdf-aa49-32419d364534.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 808.317802] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94b8de4b-d0b5-442b-9ac1-cda449090aa7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.325444] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 808.325444] env[62814]: value = "task-4293680" [ 808.325444] env[62814]: _type = "Task" [ 808.325444] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.337014] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293680, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.547998] env[62814]: DEBUG nova.scheduler.client.report [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 808.597753] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293679, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.143084} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.599174] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 808.600422] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd9f2a67-e721-4250-b8df-a6003d11788a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.632826] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 2d0b24c5-3593-4ef2-a637-d3590242ad79/2d0b24c5-3593-4ef2-a637-d3590242ad79.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 808.633690] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0110d2e0-b767-4c58-8f46-2a64496cb884 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.657920] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 808.657920] env[62814]: value = "task-4293681" [ 808.657920] env[62814]: _type = "Task" [ 808.657920] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.667427] env[62814]: DEBUG oslo_concurrency.lockutils [None req-14269995-572b-46be-9b43-7367cd2a8f8d tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lock "1f40dc62-1a58-4cfb-8785-c37b68747f37" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 86.866s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 808.668016] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293681, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.848259] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293680, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.865735] env[62814]: DEBUG nova.network.neutron [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.961838] env[62814]: DEBUG nova.network.neutron [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.055664] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.241s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.057742] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquiring lock "cd89e4d5-c08e-42c9-aea8-84c20714e196" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.057983] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lock "cd89e4d5-c08e-42c9-aea8-84c20714e196" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.058834] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.933s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.060027] env[62814]: INFO nova.compute.claims [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.090420] env[62814]: INFO nova.scheduler.client.report [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Deleted allocations for instance 64a475e4-6713-408b-a63a-a43b5fed5ec8 [ 809.137295] env[62814]: DEBUG nova.compute.manager [req-51750eed-2595-4ae2-8e3c-b0dd97518d17 req-1419539a-87f4-4f04-a74c-4ae3ef3b4739 service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Received event network-changed-f5994178-e05c-464f-afea-a4f8a606efe3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 809.137758] env[62814]: DEBUG nova.compute.manager [req-51750eed-2595-4ae2-8e3c-b0dd97518d17 req-1419539a-87f4-4f04-a74c-4ae3ef3b4739 service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Refreshing instance network info cache due to event network-changed-f5994178-e05c-464f-afea-a4f8a606efe3. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 809.139730] env[62814]: DEBUG oslo_concurrency.lockutils [req-51750eed-2595-4ae2-8e3c-b0dd97518d17 req-1419539a-87f4-4f04-a74c-4ae3ef3b4739 service nova] Acquiring lock "refresh_cache-1f40dc62-1a58-4cfb-8785-c37b68747f37" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.139730] env[62814]: DEBUG oslo_concurrency.lockutils [req-51750eed-2595-4ae2-8e3c-b0dd97518d17 req-1419539a-87f4-4f04-a74c-4ae3ef3b4739 service nova] Acquired lock "refresh_cache-1f40dc62-1a58-4cfb-8785-c37b68747f37" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.139730] env[62814]: DEBUG nova.network.neutron [req-51750eed-2595-4ae2-8e3c-b0dd97518d17 req-1419539a-87f4-4f04-a74c-4ae3ef3b4739 service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Refreshing network info cache for port f5994178-e05c-464f-afea-a4f8a606efe3 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 809.154637] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "59d21ef7-df97-49ac-9329-4c18df6dd087" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.154891] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.155163] env[62814]: INFO nova.compute.manager [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Rebooting instance [ 809.170460] env[62814]: DEBUG nova.compute.manager [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 809.172910] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293681, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.336844] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293680, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.729926} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.337130] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c18acec7-cf95-4cdf-aa49-32419d364534/c18acec7-cf95-4cdf-aa49-32419d364534.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 809.337358] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.337618] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7962e155-aaa8-43bc-8e10-742275376413 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.345453] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 809.345453] env[62814]: value = "task-4293682" [ 809.345453] env[62814]: _type = "Task" [ 809.345453] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.352923] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293682, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.468018] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Releasing lock "refresh_cache-127fee64-fd56-4a23-bdd2-18c817898fd5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 809.469095] env[62814]: DEBUG nova.compute.manager [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 809.469095] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 809.470309] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b426415-7c74-440c-a25a-4e2b78116834 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.477838] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 809.478123] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c4d5a8b-1daa-44ff-ab80-dbb21960fa3d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.484646] env[62814]: DEBUG oslo_vmware.api [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 809.484646] env[62814]: value = "task-4293683" [ 809.484646] env[62814]: _type = "Task" [ 809.484646] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.493082] env[62814]: DEBUG oslo_vmware.api [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293683, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.607471] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97cc783f-5241-4f40-9b22-e9c65674583a tempest-ListServersNegativeTestJSON-1833739842 tempest-ListServersNegativeTestJSON-1833739842-project-member] Lock "64a475e4-6713-408b-a63a-a43b5fed5ec8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.962s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.668886] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293681, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.679508] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.679681] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.679884] env[62814]: DEBUG nova.network.neutron [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.692028] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.824328] env[62814]: DEBUG oslo_concurrency.lockutils [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquiring lock "9677b03f-7138-47b7-b1e8-f3714d11e550" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.824328] env[62814]: DEBUG oslo_concurrency.lockutils [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lock "9677b03f-7138-47b7-b1e8-f3714d11e550" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.824328] env[62814]: DEBUG oslo_concurrency.lockutils [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquiring lock "9677b03f-7138-47b7-b1e8-f3714d11e550-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 809.824328] env[62814]: DEBUG oslo_concurrency.lockutils [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lock "9677b03f-7138-47b7-b1e8-f3714d11e550-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.824328] env[62814]: DEBUG oslo_concurrency.lockutils [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lock "9677b03f-7138-47b7-b1e8-f3714d11e550-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.825508] env[62814]: INFO nova.compute.manager [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Terminating instance [ 809.856349] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293682, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.197733} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.856871] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.857772] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c83a41-1c7d-4cd8-a77e-7388a0c23a7f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.882897] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] c18acec7-cf95-4cdf-aa49-32419d364534/c18acec7-cf95-4cdf-aa49-32419d364534.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.885875] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0b3a4e6-6a64-48d4-822b-37b6a8607221 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.911147] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 809.911147] env[62814]: value = "task-4293684" [ 809.911147] env[62814]: _type = "Task" [ 809.911147] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.917461] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293684, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.973374] env[62814]: DEBUG nova.network.neutron [req-51750eed-2595-4ae2-8e3c-b0dd97518d17 req-1419539a-87f4-4f04-a74c-4ae3ef3b4739 service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Updated VIF entry in instance network info cache for port f5994178-e05c-464f-afea-a4f8a606efe3. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 809.973519] env[62814]: DEBUG nova.network.neutron [req-51750eed-2595-4ae2-8e3c-b0dd97518d17 req-1419539a-87f4-4f04-a74c-4ae3ef3b4739 service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Updating instance_info_cache with network_info: [{"id": "f5994178-e05c-464f-afea-a4f8a606efe3", "address": "fa:16:3e:ba:17:3e", "network": {"id": "12e384b1-79ae-4b53-8288-2b6be23e6640", "bridge": "br-int", "label": "tempest-ServersTestJSON-1101498603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.226", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ab4d42a6e2e54eaea18742dd5fb56c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604056d6-6dd6-47fa-9eaa-6863a3a7c488", "external-id": "nsx-vlan-transportzone-287", "segmentation_id": 287, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5994178-e0", "ovs_interfaceid": "f5994178-e05c-464f-afea-a4f8a606efe3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.997298] env[62814]: DEBUG oslo_vmware.api [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293683, 'name': PowerOffVM_Task, 'duration_secs': 0.129603} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.997298] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 809.997298] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 809.997298] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a5ba4df9-adbb-4bd6-888c-9beaab68246c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.021542] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.021628] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.021748] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Deleting the datastore file [datastore2] 127fee64-fd56-4a23-bdd2-18c817898fd5 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.022036] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9eb6345b-83f1-4b63-b31e-ac9dd0f36c4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.030609] env[62814]: DEBUG oslo_vmware.api [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for the task: (returnval){ [ 810.030609] env[62814]: value = "task-4293686" [ 810.030609] env[62814]: _type = "Task" [ 810.030609] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.041026] env[62814]: DEBUG oslo_vmware.api [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293686, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.172053] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293681, 'name': ReconfigVM_Task, 'duration_secs': 1.214465} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.172360] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 2d0b24c5-3593-4ef2-a637-d3590242ad79/2d0b24c5-3593-4ef2-a637-d3590242ad79.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.173022] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5fabcfdc-b8ff-4603-8f91-7320537339ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.179624] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 810.179624] env[62814]: value = "task-4293687" [ 810.179624] env[62814]: _type = "Task" [ 810.179624] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.192716] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293687, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.330729] env[62814]: DEBUG nova.compute.manager [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 810.331060] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 810.332055] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72eb8d63-a197-46a0-aa02-af3a9d70829d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.344048] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 810.344959] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a569e928-f6d4-4208-b792-5cd78808a9ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.351531] env[62814]: DEBUG oslo_vmware.api [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 810.351531] env[62814]: value = "task-4293688" [ 810.351531] env[62814]: _type = "Task" [ 810.351531] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.368954] env[62814]: DEBUG oslo_vmware.api [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293688, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.425866] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293684, 'name': ReconfigVM_Task, 'duration_secs': 0.287288} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.428539] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Reconfigured VM instance instance-0000002f to attach disk [datastore2] c18acec7-cf95-4cdf-aa49-32419d364534/c18acec7-cf95-4cdf-aa49-32419d364534.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.430015] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b7dd1ea-d777-4d07-a6d0-3618b9b3c3d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.439081] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 810.439081] env[62814]: value = "task-4293689" [ 810.439081] env[62814]: _type = "Task" [ 810.439081] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.455137] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293689, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.480045] env[62814]: DEBUG oslo_concurrency.lockutils [req-51750eed-2595-4ae2-8e3c-b0dd97518d17 req-1419539a-87f4-4f04-a74c-4ae3ef3b4739 service nova] Releasing lock "refresh_cache-1f40dc62-1a58-4cfb-8785-c37b68747f37" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.486069] env[62814]: DEBUG nova.network.neutron [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance_info_cache with network_info: [{"id": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "address": "fa:16:3e:93:37:3e", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa00402c1-eb", "ovs_interfaceid": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.547498] env[62814]: DEBUG oslo_vmware.api [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Task: {'id': task-4293686, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.202161} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.554120] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.554120] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 810.554120] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 810.554120] env[62814]: INFO nova.compute.manager [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Took 1.08 seconds to destroy the instance on the hypervisor. [ 810.554120] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 810.554439] env[62814]: DEBUG nova.compute.manager [-] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 810.554627] env[62814]: DEBUG nova.network.neutron [-] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 810.576260] env[62814]: DEBUG nova.network.neutron [-] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.663797] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2feea82-a9ca-479f-93a2-ae97b85fcc49 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.676393] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437af73b-d78d-48b7-81b8-16de3b450a49 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.716703] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6cacaf1-768e-4347-a0dc-ce36a33b9afd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.725908] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293687, 'name': Rename_Task, 'duration_secs': 0.328981} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.729256] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.729608] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e6757b62-d955-4801-bda2-ecf5d8f6ac16 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.732772] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a73d62a1-bb14-4cb2-a180-aeac14393253 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.750719] env[62814]: DEBUG nova.compute.provider_tree [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.753860] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 810.753860] env[62814]: value = "task-4293690" [ 810.753860] env[62814]: _type = "Task" [ 810.753860] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.861857] env[62814]: DEBUG oslo_vmware.api [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293688, 'name': PowerOffVM_Task, 'duration_secs': 0.187459} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.862164] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 810.862329] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 810.862586] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-746ae061-e891-48bd-b939-38662a1abc24 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.916818] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 810.917071] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 810.917259] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Deleting the datastore file [datastore2] 9677b03f-7138-47b7-b1e8-f3714d11e550 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.917549] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-89de4839-c08d-47ea-b73b-19600b9d30f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.924772] env[62814]: DEBUG oslo_vmware.api [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for the task: (returnval){ [ 810.924772] env[62814]: value = "task-4293692" [ 810.924772] env[62814]: _type = "Task" [ 810.924772] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.932345] env[62814]: DEBUG oslo_vmware.api [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293692, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.947301] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293689, 'name': Rename_Task, 'duration_secs': 0.167922} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.947563] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.947874] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8543dcaa-7ccb-4ff0-b023-de4c61c90cd3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.954043] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 810.954043] env[62814]: value = "task-4293693" [ 810.954043] env[62814]: _type = "Task" [ 810.954043] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.961724] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293693, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.992902] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 811.079275] env[62814]: DEBUG nova.network.neutron [-] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.256319] env[62814]: DEBUG nova.scheduler.client.report [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 811.273240] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293690, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.436801] env[62814]: DEBUG oslo_vmware.api [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Task: {'id': task-4293692, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.298905} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.437226] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 811.437435] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 811.437618] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 811.437835] env[62814]: INFO nova.compute.manager [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Took 1.11 seconds to destroy the instance on the hypervisor. [ 811.438102] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 811.438315] env[62814]: DEBUG nova.compute.manager [-] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 811.438406] env[62814]: DEBUG nova.network.neutron [-] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 811.464591] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293693, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.501068] env[62814]: DEBUG nova.compute.manager [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 811.502533] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c383dd-2a29-4fb3-b353-e6b44b7df304 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.582432] env[62814]: INFO nova.compute.manager [-] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Took 1.03 seconds to deallocate network for instance. [ 811.765314] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.707s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 811.765851] env[62814]: DEBUG nova.compute.manager [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 811.769493] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.612s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 811.769826] env[62814]: DEBUG nova.objects.instance [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lazy-loading 'resources' on Instance uuid 1a16bd56-0992-4bec-bd3f-2836dfe68579 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.777396] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293690, 'name': PowerOnVM_Task, 'duration_secs': 0.712391} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.778052] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.778157] env[62814]: INFO nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Took 10.76 seconds to spawn the instance on the hypervisor. [ 811.778328] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 811.779961] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89e22eb-f8ea-48fb-ada7-f3fc8c90b6bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.937805] env[62814]: DEBUG nova.compute.manager [req-5673c4f2-2fb2-43f3-adf6-1943eae2cade req-83dc7091-24f2-4a7a-9b80-d7a597ec660f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Received event network-vif-deleted-c108416a-2aa8-4882-a9a2-d0cbfbb4db61 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 811.938301] env[62814]: INFO nova.compute.manager [req-5673c4f2-2fb2-43f3-adf6-1943eae2cade req-83dc7091-24f2-4a7a-9b80-d7a597ec660f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Neutron deleted interface c108416a-2aa8-4882-a9a2-d0cbfbb4db61; detaching it from the instance and deleting it from the info cache [ 811.938301] env[62814]: DEBUG nova.network.neutron [req-5673c4f2-2fb2-43f3-adf6-1943eae2cade req-83dc7091-24f2-4a7a-9b80-d7a597ec660f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.963869] env[62814]: DEBUG oslo_vmware.api [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293693, 'name': PowerOnVM_Task, 'duration_secs': 0.575452} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.964157] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.964355] env[62814]: INFO nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Took 8.19 seconds to spawn the instance on the hypervisor. [ 811.964530] env[62814]: DEBUG nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 811.965305] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646a2953-693e-4001-98a7-2f3cae90bb76 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.088606] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.274042] env[62814]: DEBUG nova.compute.utils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 812.275258] env[62814]: DEBUG nova.compute.manager [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 812.275430] env[62814]: DEBUG nova.network.neutron [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 812.298732] env[62814]: INFO nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Took 46.31 seconds to build instance. [ 812.357783] env[62814]: DEBUG nova.network.neutron [-] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.367662] env[62814]: DEBUG nova.policy [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd147453e63fd4439829b3502a3d6db97', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8003fcc64a3147a296244034cfd77c9f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 812.441569] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4eb450a5-c445-436d-aebd-3bba81990937 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.454016] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006bb093-e680-4046-8947-a8299c3c7947 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.499977] env[62814]: DEBUG nova.compute.manager [req-5673c4f2-2fb2-43f3-adf6-1943eae2cade req-83dc7091-24f2-4a7a-9b80-d7a597ec660f service nova] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Detach interface failed, port_id=c108416a-2aa8-4882-a9a2-d0cbfbb4db61, reason: Instance 9677b03f-7138-47b7-b1e8-f3714d11e550 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 812.502157] env[62814]: INFO nova.compute.manager [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Took 43.14 seconds to build instance. [ 812.520918] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfff943-3c21-47f8-a6dc-55676ba7ae04 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.534022] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Doing hard reboot of VM {{(pid=62814) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 812.534022] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-3e9f428a-877c-4c3a-8b7e-6980d91aa592 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.538861] env[62814]: DEBUG oslo_vmware.api [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 812.538861] env[62814]: value = "task-4293694" [ 812.538861] env[62814]: _type = "Task" [ 812.538861] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.550766] env[62814]: DEBUG oslo_vmware.api [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293694, 'name': ResetVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.780736] env[62814]: DEBUG nova.compute.manager [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 812.804188] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "2d0b24c5-3593-4ef2-a637-d3590242ad79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.405s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.828884] env[62814]: DEBUG nova.network.neutron [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Successfully created port: bc38976c-8838-47e6-b4f8-cd013e62b15b {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 812.865176] env[62814]: INFO nova.compute.manager [-] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Took 1.43 seconds to deallocate network for instance. [ 812.907243] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85aafe76-b664-4ef5-aae7-3b404dd47ec8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.917216] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353152da-ad0f-4fd4-ad86-d8c72c41638c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.951675] env[62814]: DEBUG oslo_concurrency.lockutils [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "2d0b24c5-3593-4ef2-a637-d3590242ad79" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.951867] env[62814]: DEBUG oslo_concurrency.lockutils [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "2d0b24c5-3593-4ef2-a637-d3590242ad79" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.952159] env[62814]: DEBUG oslo_concurrency.lockutils [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "2d0b24c5-3593-4ef2-a637-d3590242ad79-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 812.952406] env[62814]: DEBUG oslo_concurrency.lockutils [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "2d0b24c5-3593-4ef2-a637-d3590242ad79-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.952617] env[62814]: DEBUG oslo_concurrency.lockutils [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "2d0b24c5-3593-4ef2-a637-d3590242ad79-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.954858] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a88469a-346d-4836-b942-6df44f496db7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.959019] env[62814]: INFO nova.compute.manager [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Terminating instance [ 812.965493] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932457d0-1120-43f6-8b93-3ac90a8bf48e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.982966] env[62814]: DEBUG nova.compute.provider_tree [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.004777] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9191f36f-df92-4599-9352-b3d1b1554eb5 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "c18acec7-cf95-4cdf-aa49-32419d364534" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.561s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.048782] env[62814]: DEBUG oslo_vmware.api [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293694, 'name': ResetVM_Task, 'duration_secs': 0.114787} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.049097] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Did hard reboot of VM {{(pid=62814) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 813.049304] env[62814]: DEBUG nova.compute.manager [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 813.050071] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f9b236-643b-46d5-8025-8bea5693aad6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.090247] env[62814]: DEBUG oslo_concurrency.lockutils [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "c18acec7-cf95-4cdf-aa49-32419d364534" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.090631] env[62814]: DEBUG oslo_concurrency.lockutils [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "c18acec7-cf95-4cdf-aa49-32419d364534" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 813.093247] env[62814]: DEBUG oslo_concurrency.lockutils [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "c18acec7-cf95-4cdf-aa49-32419d364534-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.093247] env[62814]: DEBUG oslo_concurrency.lockutils [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "c18acec7-cf95-4cdf-aa49-32419d364534-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 813.093247] env[62814]: DEBUG oslo_concurrency.lockutils [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "c18acec7-cf95-4cdf-aa49-32419d364534-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.094544] env[62814]: INFO nova.compute.manager [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Terminating instance [ 813.310198] env[62814]: DEBUG nova.compute.manager [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 813.371033] env[62814]: DEBUG oslo_concurrency.lockutils [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.462135] env[62814]: DEBUG nova.compute.manager [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 813.462489] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 813.463767] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54ef9d4-7f2d-47ad-b03e-168d35cc5ed2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.473067] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 813.473900] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-de053e92-d63f-438e-af79-7faf4e39954e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.482395] env[62814]: DEBUG oslo_vmware.api [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 813.482395] env[62814]: value = "task-4293695" [ 813.482395] env[62814]: _type = "Task" [ 813.482395] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.485914] env[62814]: DEBUG nova.scheduler.client.report [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 813.494289] env[62814]: DEBUG oslo_vmware.api [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293695, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.507666] env[62814]: DEBUG nova.compute.manager [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 813.562790] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3609c21c-7feb-4f80-94c0-69e490d3625c tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.408s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.598463] env[62814]: DEBUG nova.compute.manager [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 813.598674] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 813.599700] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15cc8f01-c568-46e4-a23c-558f964367ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.609553] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 813.609880] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bcb74fde-0ca1-4f95-a2ec-a7b9d8867d3e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.616925] env[62814]: DEBUG oslo_vmware.api [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 813.616925] env[62814]: value = "task-4293696" [ 813.616925] env[62814]: _type = "Task" [ 813.616925] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.626063] env[62814]: DEBUG oslo_vmware.api [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293696, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.793065] env[62814]: DEBUG nova.compute.manager [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 813.826034] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:36:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='82139a72-836d-436a-81f9-b33cf43b83c7',id=40,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1395970101',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 813.826256] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 813.826540] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 813.826641] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 813.830110] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 813.830110] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 813.830110] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 813.830110] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 813.830110] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 813.830110] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 813.830110] env[62814]: DEBUG nova.virt.hardware [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 813.830110] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f747dd-3376-4d4c-8890-9359d408579a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.839629] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 813.840652] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-784dd916-2acd-49f0-a75e-773897b20fba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.993023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.224s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.994984] env[62814]: DEBUG oslo_vmware.api [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293695, 'name': PowerOffVM_Task, 'duration_secs': 0.212215} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.995445] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.800s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 813.995668] env[62814]: DEBUG nova.objects.instance [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lazy-loading 'resources' on Instance uuid 217e5812-95cc-4104-8d7b-82dfca2c0fcc {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 813.996883] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 813.997079] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 813.997325] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1f3cdbe8-d051-4f32-bc38-c7b15fe43f65 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.020422] env[62814]: INFO nova.scheduler.client.report [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Deleted allocations for instance 1a16bd56-0992-4bec-bd3f-2836dfe68579 [ 814.031693] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.054953] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 814.055219] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 814.055453] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleting the datastore file [datastore2] 2d0b24c5-3593-4ef2-a637-d3590242ad79 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 814.055721] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a8b8ea7b-20f1-4d90-b7f3-d8c1a66bc502 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.062446] env[62814]: DEBUG oslo_vmware.api [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 814.062446] env[62814]: value = "task-4293698" [ 814.062446] env[62814]: _type = "Task" [ 814.062446] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.071486] env[62814]: DEBUG oslo_vmware.api [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.128022] env[62814]: DEBUG oslo_vmware.api [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293696, 'name': PowerOffVM_Task, 'duration_secs': 0.209935} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.128022] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 814.128022] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 814.128022] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a770794-a227-478f-a85e-37a72280fabb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.188050] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 814.188352] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 814.188556] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleting the datastore file [datastore2] c18acec7-cf95-4cdf-aa49-32419d364534 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 814.188823] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-889d2876-3546-4855-b5c4-bea54627722c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.195999] env[62814]: DEBUG oslo_vmware.api [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for the task: (returnval){ [ 814.195999] env[62814]: value = "task-4293700" [ 814.195999] env[62814]: _type = "Task" [ 814.195999] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.205259] env[62814]: DEBUG oslo_vmware.api [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293700, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.400152] env[62814]: DEBUG nova.compute.manager [req-c183ca32-487f-417a-be40-79715a22c628 req-f75a5510-2a30-471c-ab98-d4e8d0e6b810 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Received event network-vif-plugged-bc38976c-8838-47e6-b4f8-cd013e62b15b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 814.400152] env[62814]: DEBUG oslo_concurrency.lockutils [req-c183ca32-487f-417a-be40-79715a22c628 req-f75a5510-2a30-471c-ab98-d4e8d0e6b810 service nova] Acquiring lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 814.400152] env[62814]: DEBUG oslo_concurrency.lockutils [req-c183ca32-487f-417a-be40-79715a22c628 req-f75a5510-2a30-471c-ab98-d4e8d0e6b810 service nova] Lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.400152] env[62814]: DEBUG oslo_concurrency.lockutils [req-c183ca32-487f-417a-be40-79715a22c628 req-f75a5510-2a30-471c-ab98-d4e8d0e6b810 service nova] Lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.400152] env[62814]: DEBUG nova.compute.manager [req-c183ca32-487f-417a-be40-79715a22c628 req-f75a5510-2a30-471c-ab98-d4e8d0e6b810 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] No waiting events found dispatching network-vif-plugged-bc38976c-8838-47e6-b4f8-cd013e62b15b {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 814.400152] env[62814]: WARNING nova.compute.manager [req-c183ca32-487f-417a-be40-79715a22c628 req-f75a5510-2a30-471c-ab98-d4e8d0e6b810 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Received unexpected event network-vif-plugged-bc38976c-8838-47e6-b4f8-cd013e62b15b for instance with vm_state building and task_state spawning. [ 814.509655] env[62814]: DEBUG nova.network.neutron [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Successfully updated port: bc38976c-8838-47e6-b4f8-cd013e62b15b {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 814.531137] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81201550-9b1e-420b-8a9d-bb1c5f1b1837 tempest-InstanceActionsTestJSON-953202637 tempest-InstanceActionsTestJSON-953202637-project-member] Lock "1a16bd56-0992-4bec-bd3f-2836dfe68579" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 40.789s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.574251] env[62814]: DEBUG oslo_vmware.api [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197238} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.574512] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.574730] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 814.575113] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 814.575113] env[62814]: INFO nova.compute.manager [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Took 1.11 seconds to destroy the instance on the hypervisor. [ 814.575316] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 814.575523] env[62814]: DEBUG nova.compute.manager [-] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 814.575621] env[62814]: DEBUG nova.network.neutron [-] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 814.707272] env[62814]: DEBUG oslo_vmware.api [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Task: {'id': task-4293700, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178152} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.709842] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 814.710048] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 814.710249] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 814.710492] env[62814]: INFO nova.compute.manager [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Took 1.11 seconds to destroy the instance on the hypervisor. [ 814.710738] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 814.711125] env[62814]: DEBUG nova.compute.manager [-] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 814.711226] env[62814]: DEBUG nova.network.neutron [-] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.006199] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0eb9646-31c3-43b6-940a-4043bdb233d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.013558] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.013704] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 815.013862] env[62814]: DEBUG nova.network.neutron [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 815.015980] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b71f1019-109b-452c-bdf2-91b122a27810 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.069119] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c978375f-ec71-42a1-a0f9-4cb19f544b1a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.081899] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f3c592-7407-4454-ba83-15043d0ca177 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.106878] env[62814]: DEBUG nova.compute.provider_tree [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.125729] env[62814]: DEBUG nova.compute.manager [req-112163c0-be4f-4398-9ba4-313ff92a57d1 req-fd7dfa6b-8ec6-4cad-b330-2ab5c200cb7a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Received event network-vif-deleted-d79a3b82-ece8-4a3e-9549-734213a374c7 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 815.125931] env[62814]: INFO nova.compute.manager [req-112163c0-be4f-4398-9ba4-313ff92a57d1 req-fd7dfa6b-8ec6-4cad-b330-2ab5c200cb7a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Neutron deleted interface d79a3b82-ece8-4a3e-9549-734213a374c7; detaching it from the instance and deleting it from the info cache [ 815.126174] env[62814]: DEBUG nova.network.neutron [req-112163c0-be4f-4398-9ba4-313ff92a57d1 req-fd7dfa6b-8ec6-4cad-b330-2ab5c200cb7a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.324672] env[62814]: DEBUG nova.network.neutron [-] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.558922] env[62814]: DEBUG nova.network.neutron [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.574542] env[62814]: DEBUG nova.network.neutron [-] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.612567] env[62814]: DEBUG nova.scheduler.client.report [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 815.628848] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ce9b5ae-a9d8-4151-9d9d-42f88c910b02 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.640063] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c378ed-d564-42c7-a1e2-8904557af368 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.678311] env[62814]: DEBUG nova.compute.manager [req-112163c0-be4f-4398-9ba4-313ff92a57d1 req-fd7dfa6b-8ec6-4cad-b330-2ab5c200cb7a service nova] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Detach interface failed, port_id=d79a3b82-ece8-4a3e-9549-734213a374c7, reason: Instance c18acec7-cf95-4cdf-aa49-32419d364534 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 815.717268] env[62814]: DEBUG nova.network.neutron [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance_info_cache with network_info: [{"id": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "address": "fa:16:3e:fe:5b:6c", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38976c-88", "ovs_interfaceid": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.828678] env[62814]: INFO nova.compute.manager [-] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Took 1.25 seconds to deallocate network for instance. [ 816.077450] env[62814]: INFO nova.compute.manager [-] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Took 1.37 seconds to deallocate network for instance. [ 816.117895] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.122s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.120785] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.330s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.122190] env[62814]: INFO nova.compute.claims [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 816.146618] env[62814]: INFO nova.scheduler.client.report [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted allocations for instance 217e5812-95cc-4104-8d7b-82dfca2c0fcc [ 816.219645] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 816.220038] env[62814]: DEBUG nova.compute.manager [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Instance network_info: |[{"id": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "address": "fa:16:3e:fe:5b:6c", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38976c-88", "ovs_interfaceid": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 816.220753] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:5b:6c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc38976c-8838-47e6-b4f8-cd013e62b15b', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 816.228894] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 816.229656] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 816.229656] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a354f297-abf1-4596-b8ad-2a29c1ecd2a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.253894] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 816.253894] env[62814]: value = "task-4293701" [ 816.253894] env[62814]: _type = "Task" [ 816.253894] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.263099] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293701, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.338065] env[62814]: DEBUG oslo_concurrency.lockutils [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.432621] env[62814]: DEBUG nova.compute.manager [req-f9dfde1b-2fe9-47b8-8b7b-9f25f83f3678 req-f732abe9-409a-4bfd-8efd-4755fb91c797 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Received event network-changed-bc38976c-8838-47e6-b4f8-cd013e62b15b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 816.432877] env[62814]: DEBUG nova.compute.manager [req-f9dfde1b-2fe9-47b8-8b7b-9f25f83f3678 req-f732abe9-409a-4bfd-8efd-4755fb91c797 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Refreshing instance network info cache due to event network-changed-bc38976c-8838-47e6-b4f8-cd013e62b15b. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 816.433161] env[62814]: DEBUG oslo_concurrency.lockutils [req-f9dfde1b-2fe9-47b8-8b7b-9f25f83f3678 req-f732abe9-409a-4bfd-8efd-4755fb91c797 service nova] Acquiring lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.433317] env[62814]: DEBUG oslo_concurrency.lockutils [req-f9dfde1b-2fe9-47b8-8b7b-9f25f83f3678 req-f732abe9-409a-4bfd-8efd-4755fb91c797 service nova] Acquired lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 816.433561] env[62814]: DEBUG nova.network.neutron [req-f9dfde1b-2fe9-47b8-8b7b-9f25f83f3678 req-f732abe9-409a-4bfd-8efd-4755fb91c797 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Refreshing network info cache for port bc38976c-8838-47e6-b4f8-cd013e62b15b {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 816.584068] env[62814]: DEBUG oslo_concurrency.lockutils [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.660022] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5fc46c21-e72a-49b2-8525-23eb41efd287 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "217e5812-95cc-4104-8d7b-82dfca2c0fcc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.335s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.764026] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293701, 'name': CreateVM_Task, 'duration_secs': 0.325143} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.764208] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 816.764985] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.765172] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 816.765512] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 816.765764] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d924287-0433-4d63-8b76-55c8d02a9ff8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.770719] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 816.770719] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52442a42-b378-e54f-8424-16b045a0aec7" [ 816.770719] env[62814]: _type = "Task" [ 816.770719] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.779060] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52442a42-b378-e54f-8424-16b045a0aec7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.211543] env[62814]: DEBUG nova.network.neutron [req-f9dfde1b-2fe9-47b8-8b7b-9f25f83f3678 req-f732abe9-409a-4bfd-8efd-4755fb91c797 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updated VIF entry in instance network info cache for port bc38976c-8838-47e6-b4f8-cd013e62b15b. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 817.212912] env[62814]: DEBUG nova.network.neutron [req-f9dfde1b-2fe9-47b8-8b7b-9f25f83f3678 req-f732abe9-409a-4bfd-8efd-4755fb91c797 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance_info_cache with network_info: [{"id": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "address": "fa:16:3e:fe:5b:6c", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38976c-88", "ovs_interfaceid": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.281879] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52442a42-b378-e54f-8424-16b045a0aec7, 'name': SearchDatastore_Task, 'duration_secs': 0.022508} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.284515] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 817.284747] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 817.285007] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.285166] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 817.285345] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 817.285776] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-775d1b6a-2c88-4eb9-8b1f-dc1bc6f01047 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.294999] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 817.295201] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 817.298188] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e313f5e0-436f-4175-af4f-ca8be33bfa63 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.303484] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 817.303484] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524831c4-feed-06cb-e10b-06f39be0d893" [ 817.303484] env[62814]: _type = "Task" [ 817.303484] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.311347] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524831c4-feed-06cb-e10b-06f39be0d893, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.557415] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-472a2e66-0cc6-4093-95c2-c5031b15f2bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.565743] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a34f69-3d5d-4067-ab0a-65f5847fe206 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.594578] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4780268d-8e4d-49ed-9817-b686867949ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.602257] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece60a10-c8a6-44bc-adcd-10612582e1e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.615020] env[62814]: DEBUG nova.compute.provider_tree [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.714633] env[62814]: DEBUG oslo_concurrency.lockutils [req-f9dfde1b-2fe9-47b8-8b7b-9f25f83f3678 req-f732abe9-409a-4bfd-8efd-4755fb91c797 service nova] Releasing lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 817.715183] env[62814]: DEBUG nova.compute.manager [req-f9dfde1b-2fe9-47b8-8b7b-9f25f83f3678 req-f732abe9-409a-4bfd-8efd-4755fb91c797 service nova] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Received event network-vif-deleted-079f803d-e719-4778-8920-cecd6ea5510c {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 817.815106] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524831c4-feed-06cb-e10b-06f39be0d893, 'name': SearchDatastore_Task, 'duration_secs': 0.016835} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.815106] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c521833-a4a9-468e-9525-2227d865fbac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.819376] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 817.819376] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527e9cf5-82bb-cf77-28e7-50d563470c60" [ 817.819376] env[62814]: _type = "Task" [ 817.819376] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.826924] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527e9cf5-82bb-cf77-28e7-50d563470c60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.118017] env[62814]: DEBUG nova.scheduler.client.report [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 818.333155] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527e9cf5-82bb-cf77-28e7-50d563470c60, 'name': SearchDatastore_Task, 'duration_secs': 0.009383} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.333432] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 818.333679] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 9cd89da4-d190-4bfc-81e1-da15b98d54ba/9cd89da4-d190-4bfc-81e1-da15b98d54ba.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 818.333934] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbc33ae3-b705-4d6a-8789-1ee985c84482 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.344476] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 818.344476] env[62814]: value = "task-4293702" [ 818.344476] env[62814]: _type = "Task" [ 818.344476] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.350458] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293702, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.624368] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 818.624368] env[62814]: DEBUG nova.compute.manager [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 818.626769] env[62814]: DEBUG oslo_concurrency.lockutils [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.293s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 818.627704] env[62814]: DEBUG nova.objects.instance [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lazy-loading 'resources' on Instance uuid 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 818.857114] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293702, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511402} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.857414] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 9cd89da4-d190-4bfc-81e1-da15b98d54ba/9cd89da4-d190-4bfc-81e1-da15b98d54ba.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 818.857627] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 818.857937] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e1b92c45-cd01-424f-8b90-86768ec4e922 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.869898] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 818.869898] env[62814]: value = "task-4293703" [ 818.869898] env[62814]: _type = "Task" [ 818.869898] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.880747] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293703, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.136223] env[62814]: DEBUG nova.compute.utils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 819.138699] env[62814]: DEBUG nova.compute.manager [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 819.138699] env[62814]: DEBUG nova.network.neutron [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 819.220282] env[62814]: DEBUG nova.policy [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d52f0fcc0ec42d69e6b2094145d0f28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '558682c8ea0f4887874a1763f65cb9a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 819.391169] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293703, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064612} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.395016] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.395016] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bfb8b07-1f14-43c9-9106-0fb67fde0187 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.414665] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 9cd89da4-d190-4bfc-81e1-da15b98d54ba/9cd89da4-d190-4bfc-81e1-da15b98d54ba.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.417431] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-701e7c1a-90b4-4cc3-912f-392183e92d94 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.439462] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 819.439462] env[62814]: value = "task-4293704" [ 819.439462] env[62814]: _type = "Task" [ 819.439462] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.451510] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293704, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.588324] env[62814]: DEBUG nova.network.neutron [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Successfully created port: 87aa2371-717e-4ae6-9aab-1b22613e710c {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 819.643803] env[62814]: DEBUG nova.compute.manager [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 819.717629] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578f53a6-2546-4bd6-b65a-69e7daab1234 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.729066] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81cfed57-7fba-481b-b23a-cc011b5a52ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.767106] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397aa263-40bd-4c8a-b419-0bd9bd1b4931 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.776021] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-424d447b-7b26-4906-ba43-c68c6b44049d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.787117] env[62814]: DEBUG nova.compute.provider_tree [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.956829] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293704, 'name': ReconfigVM_Task, 'duration_secs': 0.260781} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.957179] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 9cd89da4-d190-4bfc-81e1-da15b98d54ba/9cd89da4-d190-4bfc-81e1-da15b98d54ba.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 819.958619] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b636fd7d-c176-4ae2-8c9d-29cdc8c9830b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.965477] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 819.965477] env[62814]: value = "task-4293705" [ 819.965477] env[62814]: _type = "Task" [ 819.965477] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.977090] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293705, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.293584] env[62814]: DEBUG nova.scheduler.client.report [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 820.476884] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293705, 'name': Rename_Task, 'duration_secs': 0.135682} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.477232] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 820.477524] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-647339c7-4347-4103-9d32-12f5715604d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.484674] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 820.484674] env[62814]: value = "task-4293706" [ 820.484674] env[62814]: _type = "Task" [ 820.484674] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.492889] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293706, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.656948] env[62814]: DEBUG nova.compute.manager [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 820.696938] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 820.697200] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 820.697362] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 820.697541] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 820.697684] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 820.697865] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 820.702198] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 820.702431] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 820.702619] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 820.705492] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 820.705492] env[62814]: DEBUG nova.virt.hardware [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 820.705492] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35d7b4d-1242-4292-b7cc-38d95a7bf80d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.712965] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d165a7-42e6-4b3d-904b-5e282e0285b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.797877] env[62814]: DEBUG oslo_concurrency.lockutils [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.171s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 820.800487] env[62814]: DEBUG oslo_concurrency.lockutils [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 32.637s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 820.800671] env[62814]: DEBUG nova.objects.instance [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 820.825808] env[62814]: INFO nova.scheduler.client.report [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Deleted allocations for instance 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4 [ 820.997478] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293706, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.335708] env[62814]: DEBUG oslo_concurrency.lockutils [None req-30cce8ff-3cc0-4812-aa28-dbebae920bf4 tempest-ServersTestFqdnHostnames-1697884701 tempest-ServersTestFqdnHostnames-1697884701-project-member] Lock "8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.018s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.497993] env[62814]: DEBUG oslo_vmware.api [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293706, 'name': PowerOnVM_Task, 'duration_secs': 0.676618} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.503027] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 821.503027] env[62814]: INFO nova.compute.manager [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Took 7.71 seconds to spawn the instance on the hypervisor. [ 821.503027] env[62814]: DEBUG nova.compute.manager [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 821.503027] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91d82eb-ebff-4d96-85ff-9e976e72d4c1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.621966] env[62814]: DEBUG nova.compute.manager [req-9c77c9ca-ff96-4816-b6e9-959926d392fd req-4268ec70-777b-4a0d-90a0-87cbd1545452 service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Received event network-vif-plugged-87aa2371-717e-4ae6-9aab-1b22613e710c {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 821.622406] env[62814]: DEBUG oslo_concurrency.lockutils [req-9c77c9ca-ff96-4816-b6e9-959926d392fd req-4268ec70-777b-4a0d-90a0-87cbd1545452 service nova] Acquiring lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.622748] env[62814]: DEBUG oslo_concurrency.lockutils [req-9c77c9ca-ff96-4816-b6e9-959926d392fd req-4268ec70-777b-4a0d-90a0-87cbd1545452 service nova] Lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 821.622937] env[62814]: DEBUG oslo_concurrency.lockutils [req-9c77c9ca-ff96-4816-b6e9-959926d392fd req-4268ec70-777b-4a0d-90a0-87cbd1545452 service nova] Lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.623178] env[62814]: DEBUG nova.compute.manager [req-9c77c9ca-ff96-4816-b6e9-959926d392fd req-4268ec70-777b-4a0d-90a0-87cbd1545452 service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] No waiting events found dispatching network-vif-plugged-87aa2371-717e-4ae6-9aab-1b22613e710c {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 821.623524] env[62814]: WARNING nova.compute.manager [req-9c77c9ca-ff96-4816-b6e9-959926d392fd req-4268ec70-777b-4a0d-90a0-87cbd1545452 service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Received unexpected event network-vif-plugged-87aa2371-717e-4ae6-9aab-1b22613e710c for instance with vm_state building and task_state spawning. [ 821.651941] env[62814]: DEBUG nova.network.neutron [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Successfully updated port: 87aa2371-717e-4ae6-9aab-1b22613e710c {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 821.687754] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "5087b202-9bba-4489-823b-5d93cbf116e2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.688080] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "5087b202-9bba-4489-823b-5d93cbf116e2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 821.688312] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "5087b202-9bba-4489-823b-5d93cbf116e2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.688930] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "5087b202-9bba-4489-823b-5d93cbf116e2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 821.688930] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "5087b202-9bba-4489-823b-5d93cbf116e2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.691074] env[62814]: INFO nova.compute.manager [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Terminating instance [ 821.814020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-84495cc3-774e-48c8-9aca-774788601731 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.814020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.203s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 821.814020] env[62814]: DEBUG nova.objects.instance [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lazy-loading 'resources' on Instance uuid c9e11016-e92a-459e-b5ee-b0e43ce29450 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 822.027977] env[62814]: INFO nova.compute.manager [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Took 48.93 seconds to build instance. [ 822.154583] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "refresh_cache-7cca1c35-6bfc-450d-ba74-0e825b160e8f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.154731] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired lock "refresh_cache-7cca1c35-6bfc-450d-ba74-0e825b160e8f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 822.155323] env[62814]: DEBUG nova.network.neutron [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 822.197041] env[62814]: DEBUG nova.compute.manager [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 822.197041] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 822.197041] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1ed505-0039-4209-83aa-80715ea2f51e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.204512] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 822.206018] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d44daf6f-2b4d-4535-bacf-8eb291fb316c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.215020] env[62814]: DEBUG oslo_vmware.api [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 822.215020] env[62814]: value = "task-4293707" [ 822.215020] env[62814]: _type = "Task" [ 822.215020] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.223643] env[62814]: DEBUG oslo_vmware.api [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293707, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.521223] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "9be62576-5a05-473f-befd-b33f5fde9185" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 822.521877] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "9be62576-5a05-473f-befd-b33f5fde9185" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 822.530488] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2d7b966-8b2f-4646-8386-a7b79119e2b0 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.379s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 822.713634] env[62814]: DEBUG nova.network.neutron [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.727441] env[62814]: DEBUG oslo_vmware.api [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293707, 'name': PowerOffVM_Task, 'duration_secs': 0.364895} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.727709] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 822.727874] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 822.728218] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a03c39bd-4b70-4064-8cb8-e8b75eda3f94 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.790543] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 822.790763] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 822.791682] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Deleting the datastore file [datastore2] 5087b202-9bba-4489-823b-5d93cbf116e2 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 822.791682] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e9ca59d4-c435-43fc-aa97-11f1d42ee866 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.799797] env[62814]: DEBUG oslo_vmware.api [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 822.799797] env[62814]: value = "task-4293709" [ 822.799797] env[62814]: _type = "Task" [ 822.799797] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.812556] env[62814]: DEBUG oslo_vmware.api [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293709, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.873160] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f35ea2-f2d4-4d69-8275-5225b050f62c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.883872] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c72c66-6986-4253-b11a-a353af759421 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.917475] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6cb141e-5114-4aeb-b8cb-7acfad61c872 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.924714] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928b6504-1ddf-4c80-9540-e3c8406e6a6c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.672583] env[62814]: DEBUG nova.compute.manager [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 823.676988] env[62814]: DEBUG nova.compute.provider_tree [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.681739] env[62814]: DEBUG nova.compute.manager [req-536cca68-b01a-4727-891c-b4a03ff24a64 req-986ffbfb-4991-4d60-9faf-897c501b4ac5 service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Received event network-changed-87aa2371-717e-4ae6-9aab-1b22613e710c {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 823.681958] env[62814]: DEBUG nova.compute.manager [req-536cca68-b01a-4727-891c-b4a03ff24a64 req-986ffbfb-4991-4d60-9faf-897c501b4ac5 service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Refreshing instance network info cache due to event network-changed-87aa2371-717e-4ae6-9aab-1b22613e710c. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 823.682407] env[62814]: DEBUG oslo_concurrency.lockutils [req-536cca68-b01a-4727-891c-b4a03ff24a64 req-986ffbfb-4991-4d60-9faf-897c501b4ac5 service nova] Acquiring lock "refresh_cache-7cca1c35-6bfc-450d-ba74-0e825b160e8f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.693783] env[62814]: DEBUG oslo_vmware.api [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293709, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150653} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.694206] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 823.694269] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 823.694417] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.694644] env[62814]: INFO nova.compute.manager [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Took 1.50 seconds to destroy the instance on the hypervisor. [ 823.694887] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 823.695244] env[62814]: DEBUG nova.compute.manager [-] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 823.695244] env[62814]: DEBUG nova.network.neutron [-] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.751547] env[62814]: DEBUG nova.network.neutron [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Updating instance_info_cache with network_info: [{"id": "87aa2371-717e-4ae6-9aab-1b22613e710c", "address": "fa:16:3e:ea:ce:8c", "network": {"id": "88bc6d73-6bc2-47d0-8e78-633cbe326db2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-765790142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558682c8ea0f4887874a1763f65cb9a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87aa2371-71", "ovs_interfaceid": "87aa2371-717e-4ae6-9aab-1b22613e710c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.181257] env[62814]: DEBUG nova.scheduler.client.report [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 824.217164] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 824.254805] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Releasing lock "refresh_cache-7cca1c35-6bfc-450d-ba74-0e825b160e8f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 824.254805] env[62814]: DEBUG nova.compute.manager [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Instance network_info: |[{"id": "87aa2371-717e-4ae6-9aab-1b22613e710c", "address": "fa:16:3e:ea:ce:8c", "network": {"id": "88bc6d73-6bc2-47d0-8e78-633cbe326db2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-765790142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558682c8ea0f4887874a1763f65cb9a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87aa2371-71", "ovs_interfaceid": "87aa2371-717e-4ae6-9aab-1b22613e710c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 824.254805] env[62814]: DEBUG oslo_concurrency.lockutils [req-536cca68-b01a-4727-891c-b4a03ff24a64 req-986ffbfb-4991-4d60-9faf-897c501b4ac5 service nova] Acquired lock "refresh_cache-7cca1c35-6bfc-450d-ba74-0e825b160e8f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 824.254805] env[62814]: DEBUG nova.network.neutron [req-536cca68-b01a-4727-891c-b4a03ff24a64 req-986ffbfb-4991-4d60-9faf-897c501b4ac5 service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Refreshing network info cache for port 87aa2371-717e-4ae6-9aab-1b22613e710c {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.255944] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:ce:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '209639b9-c313-4b35-86dc-dccd744d174a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '87aa2371-717e-4ae6-9aab-1b22613e710c', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.273192] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 824.276212] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 824.277108] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d7dba3f-5c43-437e-bc69-f2a3808429fb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.302772] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 824.302772] env[62814]: value = "task-4293710" [ 824.302772] env[62814]: _type = "Task" [ 824.302772] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.311579] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293710, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.336758] env[62814]: DEBUG nova.compute.manager [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Stashing vm_state: active {{(pid=62814) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 824.529080] env[62814]: DEBUG nova.compute.manager [req-6e091bc9-cb47-4e19-84c3-9fcb46d799ee req-0dc7c775-31ec-4f3c-8f4b-ba2e33747a04 service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Received event network-vif-deleted-80990e09-7766-4c2a-937b-52c703caa252 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 824.529306] env[62814]: INFO nova.compute.manager [req-6e091bc9-cb47-4e19-84c3-9fcb46d799ee req-0dc7c775-31ec-4f3c-8f4b-ba2e33747a04 service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Neutron deleted interface 80990e09-7766-4c2a-937b-52c703caa252; detaching it from the instance and deleting it from the info cache [ 824.529484] env[62814]: DEBUG nova.network.neutron [req-6e091bc9-cb47-4e19-84c3-9fcb46d799ee req-0dc7c775-31ec-4f3c-8f4b-ba2e33747a04 service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.601755] env[62814]: DEBUG nova.network.neutron [req-536cca68-b01a-4727-891c-b4a03ff24a64 req-986ffbfb-4991-4d60-9faf-897c501b4ac5 service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Updated VIF entry in instance network info cache for port 87aa2371-717e-4ae6-9aab-1b22613e710c. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 824.602153] env[62814]: DEBUG nova.network.neutron [req-536cca68-b01a-4727-891c-b4a03ff24a64 req-986ffbfb-4991-4d60-9faf-897c501b4ac5 service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Updating instance_info_cache with network_info: [{"id": "87aa2371-717e-4ae6-9aab-1b22613e710c", "address": "fa:16:3e:ea:ce:8c", "network": {"id": "88bc6d73-6bc2-47d0-8e78-633cbe326db2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-765790142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558682c8ea0f4887874a1763f65cb9a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap87aa2371-71", "ovs_interfaceid": "87aa2371-717e-4ae6-9aab-1b22613e710c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.686566] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.873s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 824.689251] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.677s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 824.691034] env[62814]: INFO nova.compute.claims [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.726506] env[62814]: INFO nova.scheduler.client.report [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Deleted allocations for instance c9e11016-e92a-459e-b5ee-b0e43ce29450 [ 824.813942] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293710, 'name': CreateVM_Task, 'duration_secs': 0.344491} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.814417] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 824.814955] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.815185] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 824.815544] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 824.815836] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f90a44f-3fc9-4bcf-9d96-0fc12584f5aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.821055] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 824.821055] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fc106b-aa38-f29e-f1fa-3ed618231669" [ 824.821055] env[62814]: _type = "Task" [ 824.821055] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.829397] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fc106b-aa38-f29e-f1fa-3ed618231669, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.858301] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 824.975120] env[62814]: DEBUG nova.network.neutron [-] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.032130] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-adc17c11-9e7b-45f6-b5a8-d2ce72eef661 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.042647] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5a3e9b-7424-4181-9380-d9efe8a25c07 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.075939] env[62814]: DEBUG nova.compute.manager [req-6e091bc9-cb47-4e19-84c3-9fcb46d799ee req-0dc7c775-31ec-4f3c-8f4b-ba2e33747a04 service nova] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Detach interface failed, port_id=80990e09-7766-4c2a-937b-52c703caa252, reason: Instance 5087b202-9bba-4489-823b-5d93cbf116e2 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 825.107284] env[62814]: DEBUG oslo_concurrency.lockutils [req-536cca68-b01a-4727-891c-b4a03ff24a64 req-986ffbfb-4991-4d60-9faf-897c501b4ac5 service nova] Releasing lock "refresh_cache-7cca1c35-6bfc-450d-ba74-0e825b160e8f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 825.236292] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8652e60-b8f8-4454-97ca-700682122f34 tempest-ServerShowV257Test-1189668241 tempest-ServerShowV257Test-1189668241-project-member] Lock "c9e11016-e92a-459e-b5ee-b0e43ce29450" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.887s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 825.333203] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fc106b-aa38-f29e-f1fa-3ed618231669, 'name': SearchDatastore_Task, 'duration_secs': 0.010217} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.333520] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 825.333751] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.333993] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.334567] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 825.334567] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.334688] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-68386104-86f8-4d3f-a15b-f091bf7c90ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.343261] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.343362] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.344103] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bf23515-8103-4511-82cd-96e74b0caf6f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.349713] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 825.349713] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5291b56d-954e-bc39-7d2c-8a5351b1f174" [ 825.349713] env[62814]: _type = "Task" [ 825.349713] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.361557] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5291b56d-954e-bc39-7d2c-8a5351b1f174, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.477623] env[62814]: INFO nova.compute.manager [-] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Took 1.78 seconds to deallocate network for instance. [ 825.863572] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5291b56d-954e-bc39-7d2c-8a5351b1f174, 'name': SearchDatastore_Task, 'duration_secs': 0.008863} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.864557] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d4389f9-4dae-4b2d-bab1-6af9a6b0bf59 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.872549] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 825.872549] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dd5689-7d30-ec7d-e12d-6ba8d4c00ab8" [ 825.872549] env[62814]: _type = "Task" [ 825.872549] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.883096] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dd5689-7d30-ec7d-e12d-6ba8d4c00ab8, 'name': SearchDatastore_Task, 'duration_secs': 0.009542} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.883374] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 825.883686] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 7cca1c35-6bfc-450d-ba74-0e825b160e8f/7cca1c35-6bfc-450d-ba74-0e825b160e8f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 825.883919] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac6cc8c1-fb08-4de2-8ab9-ab3c8b34b1b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.893352] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 825.893352] env[62814]: value = "task-4293711" [ 825.893352] env[62814]: _type = "Task" [ 825.893352] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.903348] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293711, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.986213] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.197457] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0fbcbf5-aea3-41f4-a442-1a8268202562 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.207040] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-424642d0-e7cd-428e-b372-792a7313de27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.242082] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11de749-5d0b-479e-a2b4-8fca75032b64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.250148] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b5bcf7-37c4-46c3-ae50-fbe8ba137682 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.266566] env[62814]: DEBUG nova.compute.provider_tree [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 826.403265] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293711, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487955} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.403963] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 7cca1c35-6bfc-450d-ba74-0e825b160e8f/7cca1c35-6bfc-450d-ba74-0e825b160e8f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 826.403963] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 826.403963] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-afdc6c96-6db4-43fa-86a5-06bc9e88ee5b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.412188] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 826.412188] env[62814]: value = "task-4293712" [ 826.412188] env[62814]: _type = "Task" [ 826.412188] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.418704] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293712, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.793421] env[62814]: ERROR nova.scheduler.client.report [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [req-77e509b9-fbe6-4f3c-ade1-7380ea2facfe] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-77e509b9-fbe6-4f3c-ade1-7380ea2facfe"}]} [ 826.812987] env[62814]: DEBUG nova.scheduler.client.report [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 826.830629] env[62814]: DEBUG nova.scheduler.client.report [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 826.830845] env[62814]: DEBUG nova.compute.provider_tree [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 826.860635] env[62814]: DEBUG nova.scheduler.client.report [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 826.903413] env[62814]: DEBUG nova.scheduler.client.report [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 826.920156] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293712, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065306} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.922863] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 826.922863] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d874d7e-b4c5-43b4-970b-a9608e80201a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.943642] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 7cca1c35-6bfc-450d-ba74-0e825b160e8f/7cca1c35-6bfc-450d-ba74-0e825b160e8f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.946779] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7310ef5f-fc79-4cd5-bfa5-6aa425a38e13 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.971024] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 826.971024] env[62814]: value = "task-4293713" [ 826.971024] env[62814]: _type = "Task" [ 826.971024] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.976532] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293713, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.442636] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc597ca1-f35d-436a-a15b-4937292c1e8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.450786] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d199e6e9-7191-46f1-b892-2f611b606bc0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.488133] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f2e075-d750-413c-84ac-5dd363497a2c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.495818] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293713, 'name': ReconfigVM_Task, 'duration_secs': 0.29905} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.497783] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 7cca1c35-6bfc-450d-ba74-0e825b160e8f/7cca1c35-6bfc-450d-ba74-0e825b160e8f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.498425] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2b8b889c-e5cb-42a8-b7b8-5d7630256070 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.500936] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf63bbee-38c8-4de2-b915-d1040b56bce2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.514570] env[62814]: DEBUG nova.compute.provider_tree [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 827.517857] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 827.517857] env[62814]: value = "task-4293714" [ 827.517857] env[62814]: _type = "Task" [ 827.517857] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.527012] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293714, 'name': Rename_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.028980] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293714, 'name': Rename_Task, 'duration_secs': 0.13856} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.030067] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 828.030330] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edbf2909-a80b-4325-a601-a7b5703f5b01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.038225] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 828.038225] env[62814]: value = "task-4293715" [ 828.038225] env[62814]: _type = "Task" [ 828.038225] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.047605] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293715, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.095844] env[62814]: DEBUG nova.scheduler.client.report [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 78 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 828.096153] env[62814]: DEBUG nova.compute.provider_tree [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 78 to 79 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 828.096355] env[62814]: DEBUG nova.compute.provider_tree [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 828.553119] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293715, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.608226] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.916s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.608226] env[62814]: DEBUG nova.compute.manager [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 828.616262] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.010s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.616262] env[62814]: INFO nova.compute.claims [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.049269] env[62814]: DEBUG oslo_vmware.api [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293715, 'name': PowerOnVM_Task, 'duration_secs': 0.757103} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.049606] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 829.049855] env[62814]: INFO nova.compute.manager [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Took 8.39 seconds to spawn the instance on the hypervisor. [ 829.050124] env[62814]: DEBUG nova.compute.manager [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 829.050989] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3218c7fc-a27b-4592-a780-d2403f96ed80 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.115313] env[62814]: DEBUG nova.compute.utils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 829.116934] env[62814]: DEBUG nova.compute.manager [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 829.117589] env[62814]: DEBUG nova.network.neutron [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 829.180802] env[62814]: DEBUG nova.policy [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '66799e777ca94480b13359bcb0bf60cc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8b744d10fdfe4450be8aebb8540c2414', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 829.573690] env[62814]: INFO nova.compute.manager [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Took 50.81 seconds to build instance. [ 829.612799] env[62814]: DEBUG nova.network.neutron [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Successfully created port: a5c9c9ce-5548-4870-b39b-8d82d5a6bff4 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.620805] env[62814]: DEBUG nova.compute.manager [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 829.955311] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquiring lock "844737d8-d852-44bb-bf9d-e673e737ef33" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.955556] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lock "844737d8-d852-44bb-bf9d-e673e737ef33" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.078684] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34806b58-efde-494b-854e-45f530f98ac0 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 72.481s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.179990] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b08e96a-f812-4ea0-8fc4-0409e7817a99 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.188138] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c54b89-7434-4065-8f10-5f8c96012a61 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.220046] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd94451-051d-42ac-b6c2-ce6888f7b9a3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.228048] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe712c2-e8df-4837-9842-fd6a115cd24c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.245578] env[62814]: DEBUG nova.compute.provider_tree [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.587098] env[62814]: DEBUG nova.compute.manager [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 830.630446] env[62814]: DEBUG nova.compute.manager [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 830.660189] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 830.661152] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.661152] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 830.661152] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.661152] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 830.661152] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 830.661816] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 830.661816] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 830.661816] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 830.661929] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 830.662041] env[62814]: DEBUG nova.virt.hardware [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 830.663116] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0546ce0c-587c-4325-b449-bd06ecec163f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.674284] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-192dcef4-47c7-4b72-a03b-bc952dc25a6b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.753112] env[62814]: DEBUG nova.scheduler.client.report [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 831.119731] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.228597] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "1e259ec6-d31a-453b-87e0-baa446665d56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.228597] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 831.258467] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.645s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 831.259115] env[62814]: DEBUG nova.compute.manager [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 831.262181] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 34.569s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 831.263625] env[62814]: INFO nova.compute.claims [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 831.272555] env[62814]: DEBUG nova.compute.manager [req-3fbf1266-3e49-48c1-a5e3-b0ea1e9c552c req-d371d656-1a04-4cb2-bb8e-126f3abcded5 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Received event network-vif-plugged-a5c9c9ce-5548-4870-b39b-8d82d5a6bff4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 831.273089] env[62814]: DEBUG oslo_concurrency.lockutils [req-3fbf1266-3e49-48c1-a5e3-b0ea1e9c552c req-d371d656-1a04-4cb2-bb8e-126f3abcded5 service nova] Acquiring lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 831.273374] env[62814]: DEBUG oslo_concurrency.lockutils [req-3fbf1266-3e49-48c1-a5e3-b0ea1e9c552c req-d371d656-1a04-4cb2-bb8e-126f3abcded5 service nova] Lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 831.273596] env[62814]: DEBUG oslo_concurrency.lockutils [req-3fbf1266-3e49-48c1-a5e3-b0ea1e9c552c req-d371d656-1a04-4cb2-bb8e-126f3abcded5 service nova] Lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 831.273878] env[62814]: DEBUG nova.compute.manager [req-3fbf1266-3e49-48c1-a5e3-b0ea1e9c552c req-d371d656-1a04-4cb2-bb8e-126f3abcded5 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] No waiting events found dispatching network-vif-plugged-a5c9c9ce-5548-4870-b39b-8d82d5a6bff4 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 831.274032] env[62814]: WARNING nova.compute.manager [req-3fbf1266-3e49-48c1-a5e3-b0ea1e9c552c req-d371d656-1a04-4cb2-bb8e-126f3abcded5 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Received unexpected event network-vif-plugged-a5c9c9ce-5548-4870-b39b-8d82d5a6bff4 for instance with vm_state building and task_state spawning. [ 831.587719] env[62814]: DEBUG nova.network.neutron [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Successfully updated port: a5c9c9ce-5548-4870-b39b-8d82d5a6bff4 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.773138] env[62814]: DEBUG nova.compute.utils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 831.777485] env[62814]: DEBUG nova.compute.manager [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 831.777485] env[62814]: DEBUG nova.network.neutron [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 831.831439] env[62814]: DEBUG nova.policy [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67e95624a1234848a158decc960c5ef8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef503fd02adc475fa79dda8b379b6514', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 832.033110] env[62814]: DEBUG nova.compute.manager [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 832.034119] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96e93d56-6c5a-47cb-a05c-4c49114700fb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.090574] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquiring lock "refresh_cache-413eaa2e-7bbc-402e-b0d3-f030b6395d7b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.090728] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquired lock "refresh_cache-413eaa2e-7bbc-402e-b0d3-f030b6395d7b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 832.090885] env[62814]: DEBUG nova.network.neutron [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.218371] env[62814]: DEBUG nova.network.neutron [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Successfully created port: 69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.279327] env[62814]: DEBUG nova.compute.manager [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 832.544989] env[62814]: INFO nova.compute.manager [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] instance snapshotting [ 832.547694] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0538de10-811a-4fa1-9551-14bd31842a08 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.571227] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e0d13c-443e-4702-8310-935f3a791170 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.648565] env[62814]: DEBUG nova.network.neutron [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.722498] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6cc0f0-caa6-4026-b073-ac7129d402b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.737055] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b8fa62-0b3a-46ef-9afa-adab5354ea09 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.777914] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ab45a2-2a1d-42bc-9022-5b9c48b9398c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.787225] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f410abf-585f-449f-a922-d2c96b87933e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.807670] env[62814]: DEBUG nova.compute.provider_tree [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.849449] env[62814]: DEBUG nova.network.neutron [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Updating instance_info_cache with network_info: [{"id": "a5c9c9ce-5548-4870-b39b-8d82d5a6bff4", "address": "fa:16:3e:f1:07:98", "network": {"id": "173de219-d802-4163-b8ab-153d48b2f68b", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-986986852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b744d10fdfe4450be8aebb8540c2414", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cc448a80-6318-4b6a-b0a0-85fe6cc645df", "external-id": "nsx-vlan-transportzone-91", "segmentation_id": 91, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5c9c9ce-55", "ovs_interfaceid": "a5c9c9ce-5548-4870-b39b-8d82d5a6bff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.081685] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 833.082112] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1c45edce-c2ec-4f42-8849-a6e438df4d70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.089893] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 833.089893] env[62814]: value = "task-4293716" [ 833.089893] env[62814]: _type = "Task" [ 833.089893] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.097669] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293716, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.208124] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Acquiring lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.208421] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.296440] env[62814]: DEBUG nova.compute.manager [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 833.310461] env[62814]: DEBUG nova.scheduler.client.report [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 833.322517] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 833.322687] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.322880] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 833.323422] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.323689] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 833.323994] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 833.324372] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 833.324598] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 833.324835] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 833.325077] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 833.325317] env[62814]: DEBUG nova.virt.hardware [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 833.326811] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f003ab6d-59c2-48b7-91bb-ddd05bfff3be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.337552] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d18019-56ad-4921-aaa3-a658e2606ee7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.354792] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Releasing lock "refresh_cache-413eaa2e-7bbc-402e-b0d3-f030b6395d7b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 833.355191] env[62814]: DEBUG nova.compute.manager [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Instance network_info: |[{"id": "a5c9c9ce-5548-4870-b39b-8d82d5a6bff4", "address": "fa:16:3e:f1:07:98", "network": {"id": "173de219-d802-4163-b8ab-153d48b2f68b", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-986986852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b744d10fdfe4450be8aebb8540c2414", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cc448a80-6318-4b6a-b0a0-85fe6cc645df", "external-id": "nsx-vlan-transportzone-91", "segmentation_id": 91, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5c9c9ce-55", "ovs_interfaceid": "a5c9c9ce-5548-4870-b39b-8d82d5a6bff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 833.355830] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:07:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cc448a80-6318-4b6a-b0a0-85fe6cc645df', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5c9c9ce-5548-4870-b39b-8d82d5a6bff4', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.363479] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Creating folder: Project (8b744d10fdfe4450be8aebb8540c2414). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.363739] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bc27ac7c-b8a4-4675-8482-3aa91b29b552 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.367357] env[62814]: DEBUG nova.compute.manager [req-94e5ff94-e72b-40fc-bfe6-e8375b0f246b req-8cddddef-f7c0-40fb-9498-ed4e8bbdda68 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Received event network-changed-a5c9c9ce-5548-4870-b39b-8d82d5a6bff4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 833.367587] env[62814]: DEBUG nova.compute.manager [req-94e5ff94-e72b-40fc-bfe6-e8375b0f246b req-8cddddef-f7c0-40fb-9498-ed4e8bbdda68 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Refreshing instance network info cache due to event network-changed-a5c9c9ce-5548-4870-b39b-8d82d5a6bff4. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 833.367818] env[62814]: DEBUG oslo_concurrency.lockutils [req-94e5ff94-e72b-40fc-bfe6-e8375b0f246b req-8cddddef-f7c0-40fb-9498-ed4e8bbdda68 service nova] Acquiring lock "refresh_cache-413eaa2e-7bbc-402e-b0d3-f030b6395d7b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.368045] env[62814]: DEBUG oslo_concurrency.lockutils [req-94e5ff94-e72b-40fc-bfe6-e8375b0f246b req-8cddddef-f7c0-40fb-9498-ed4e8bbdda68 service nova] Acquired lock "refresh_cache-413eaa2e-7bbc-402e-b0d3-f030b6395d7b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 833.368158] env[62814]: DEBUG nova.network.neutron [req-94e5ff94-e72b-40fc-bfe6-e8375b0f246b req-8cddddef-f7c0-40fb-9498-ed4e8bbdda68 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Refreshing network info cache for port a5c9c9ce-5548-4870-b39b-8d82d5a6bff4 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.378621] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Created folder: Project (8b744d10fdfe4450be8aebb8540c2414) in parent group-v845547. [ 833.378824] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Creating folder: Instances. Parent ref: group-v845683. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.379066] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c9c17fa-338f-4755-9db8-497dad61ece9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.389683] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Created folder: Instances in parent group-v845683. [ 833.389913] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 833.390151] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.390372] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aa7d0fc3-c958-42c3-986d-e73a9f1271f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.412167] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.412167] env[62814]: value = "task-4293719" [ 833.412167] env[62814]: _type = "Task" [ 833.412167] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.419910] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293719, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.599797] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293716, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.713756] env[62814]: DEBUG nova.compute.utils [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 833.742641] env[62814]: DEBUG nova.network.neutron [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Successfully updated port: 69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 833.817803] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.556s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 833.818384] env[62814]: DEBUG nova.compute.manager [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 833.821013] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.637s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 833.821219] env[62814]: DEBUG nova.objects.instance [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lazy-loading 'resources' on Instance uuid b289800f-49ce-49e7-b6bc-a3b4ec84b434 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.926677] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293719, 'name': CreateVM_Task, 'duration_secs': 0.36485} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.926792] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 833.927530] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.927800] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 833.928337] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 833.928417] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0933f77d-9650-497d-9922-563572b951cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.933038] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for the task: (returnval){ [ 833.933038] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529db2fd-fe22-f4ca-b6fe-edc99a0b40ed" [ 833.933038] env[62814]: _type = "Task" [ 833.933038] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.940941] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529db2fd-fe22-f4ca-b6fe-edc99a0b40ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.101662] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293716, 'name': CreateSnapshot_Task, 'duration_secs': 0.518135} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.104188] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 834.104938] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b3095da-5f24-4de1-a851-9ee8ba9523a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.152010] env[62814]: DEBUG nova.network.neutron [req-94e5ff94-e72b-40fc-bfe6-e8375b0f246b req-8cddddef-f7c0-40fb-9498-ed4e8bbdda68 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Updated VIF entry in instance network info cache for port a5c9c9ce-5548-4870-b39b-8d82d5a6bff4. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 834.152427] env[62814]: DEBUG nova.network.neutron [req-94e5ff94-e72b-40fc-bfe6-e8375b0f246b req-8cddddef-f7c0-40fb-9498-ed4e8bbdda68 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Updating instance_info_cache with network_info: [{"id": "a5c9c9ce-5548-4870-b39b-8d82d5a6bff4", "address": "fa:16:3e:f1:07:98", "network": {"id": "173de219-d802-4163-b8ab-153d48b2f68b", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-986986852-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8b744d10fdfe4450be8aebb8540c2414", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cc448a80-6318-4b6a-b0a0-85fe6cc645df", "external-id": "nsx-vlan-transportzone-91", "segmentation_id": 91, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5c9c9ce-55", "ovs_interfaceid": "a5c9c9ce-5548-4870-b39b-8d82d5a6bff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.220789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 834.246189] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.246189] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquired lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 834.246189] env[62814]: DEBUG nova.network.neutron [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.324489] env[62814]: DEBUG nova.compute.utils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 834.329282] env[62814]: DEBUG nova.compute.manager [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 834.329707] env[62814]: DEBUG nova.network.neutron [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 834.383900] env[62814]: DEBUG nova.policy [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '58359539fd4545cdbf200ec364e43834', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dedbeff832a4ac48b0aa01bd6acc3f1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 834.447050] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529db2fd-fe22-f4ca-b6fe-edc99a0b40ed, 'name': SearchDatastore_Task, 'duration_secs': 0.016422} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.447050] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 834.447050] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.447050] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.447050] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 834.447050] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.449317] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77977fab-a1b6-4b94-9348-3acc5fae9320 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.457463] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.457638] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.461763] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20385c2f-6c78-4fb2-8a60-45200e65778c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.468834] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for the task: (returnval){ [ 834.468834] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5264ef4f-52a8-af6d-66f2-346d1f073e04" [ 834.468834] env[62814]: _type = "Task" [ 834.468834] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.476870] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5264ef4f-52a8-af6d-66f2-346d1f073e04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.627439] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 834.633018] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a4fab367-c5fc-4702-8b14-37fd616e09ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.644159] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 834.644159] env[62814]: value = "task-4293720" [ 834.644159] env[62814]: _type = "Task" [ 834.644159] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.656554] env[62814]: DEBUG oslo_concurrency.lockutils [req-94e5ff94-e72b-40fc-bfe6-e8375b0f246b req-8cddddef-f7c0-40fb-9498-ed4e8bbdda68 service nova] Releasing lock "refresh_cache-413eaa2e-7bbc-402e-b0d3-f030b6395d7b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 834.661646] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293720, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.746308] env[62814]: DEBUG nova.network.neutron [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Successfully created port: fa04dc3a-1732-4479-80e7-1d3a547ea8c5 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.788913] env[62814]: DEBUG nova.network.neutron [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.830531] env[62814]: DEBUG nova.compute.manager [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 834.958698] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ef02c3-4842-41b3-9132-db47d23e1bf1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.968199] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9209bb71-5573-488f-b347-e17c2800faa2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.979510] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5264ef4f-52a8-af6d-66f2-346d1f073e04, 'name': SearchDatastore_Task, 'duration_secs': 0.010413} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.008084] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9260e477-9b40-4586-9ab4-9c3f8c74669e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.010905] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a4c07e-a00e-4247-993b-a7ff0349d72d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.019745] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ac687a-f7a2-4021-be9f-e3fd727345e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.023813] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for the task: (returnval){ [ 835.023813] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aebfa6-927f-7c04-1679-fd7398b58f0e" [ 835.023813] env[62814]: _type = "Task" [ 835.023813] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.034467] env[62814]: DEBUG nova.compute.provider_tree [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.041592] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aebfa6-927f-7c04-1679-fd7398b58f0e, 'name': SearchDatastore_Task, 'duration_secs': 0.01015} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.041834] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 835.042097] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 413eaa2e-7bbc-402e-b0d3-f030b6395d7b/413eaa2e-7bbc-402e-b0d3-f030b6395d7b.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 835.042372] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-61d45c31-82c5-47ee-9202-6a93c281d707 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.046033] env[62814]: DEBUG nova.network.neutron [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Updating instance_info_cache with network_info: [{"id": "69314b3d-cb79-40f5-8562-afed90be617f", "address": "fa:16:3e:fa:61:e3", "network": {"id": "4ddbe620-bc9d-4b38-a414-7799c4e7060d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-334880125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef503fd02adc475fa79dda8b379b6514", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c964bdc6-fccc-40d9-bfe2-763b6f05a863", "external-id": "cl2-zone-376", "segmentation_id": 376, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69314b3d-cb", "ovs_interfaceid": "69314b3d-cb79-40f5-8562-afed90be617f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.053215] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for the task: (returnval){ [ 835.053215] env[62814]: value = "task-4293721" [ 835.053215] env[62814]: _type = "Task" [ 835.053215] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.064098] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293721, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.158857] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293720, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.353589] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Acquiring lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.354018] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.354018] env[62814]: INFO nova.compute.manager [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Attaching volume 996a4d77-b6db-42c8-93dc-c289397a5dd4 to /dev/sdb [ 835.399758] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f21d16-d977-4c2e-8d80-bf122cd2de5b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.405316] env[62814]: DEBUG nova.compute.manager [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Received event network-vif-plugged-69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 835.405617] env[62814]: DEBUG oslo_concurrency.lockutils [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] Acquiring lock "425e8edd-c002-45a1-bb6f-ee3ac8812509-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.405871] env[62814]: DEBUG oslo_concurrency.lockutils [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] Lock "425e8edd-c002-45a1-bb6f-ee3ac8812509-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.406098] env[62814]: DEBUG oslo_concurrency.lockutils [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] Lock "425e8edd-c002-45a1-bb6f-ee3ac8812509-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.406312] env[62814]: DEBUG nova.compute.manager [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] No waiting events found dispatching network-vif-plugged-69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 835.406938] env[62814]: WARNING nova.compute.manager [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Received unexpected event network-vif-plugged-69314b3d-cb79-40f5-8562-afed90be617f for instance with vm_state building and task_state spawning. [ 835.406938] env[62814]: DEBUG nova.compute.manager [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Received event network-changed-69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 835.406938] env[62814]: DEBUG nova.compute.manager [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Refreshing instance network info cache due to event network-changed-69314b3d-cb79-40f5-8562-afed90be617f. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 835.407169] env[62814]: DEBUG oslo_concurrency.lockutils [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] Acquiring lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.412874] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4390f411-b7bf-4e24-aa98-95fcd1314d5e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.430419] env[62814]: DEBUG nova.virt.block_device [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Updating existing volume attachment record: 782cf720-f3f7-4ef9-bd62-d64b8d076f1b {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 835.537545] env[62814]: DEBUG nova.scheduler.client.report [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 835.547919] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Releasing lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 835.548291] env[62814]: DEBUG nova.compute.manager [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Instance network_info: |[{"id": "69314b3d-cb79-40f5-8562-afed90be617f", "address": "fa:16:3e:fa:61:e3", "network": {"id": "4ddbe620-bc9d-4b38-a414-7799c4e7060d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-334880125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef503fd02adc475fa79dda8b379b6514", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c964bdc6-fccc-40d9-bfe2-763b6f05a863", "external-id": "cl2-zone-376", "segmentation_id": 376, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69314b3d-cb", "ovs_interfaceid": "69314b3d-cb79-40f5-8562-afed90be617f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 835.548585] env[62814]: DEBUG oslo_concurrency.lockutils [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] Acquired lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 835.548770] env[62814]: DEBUG nova.network.neutron [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Refreshing network info cache for port 69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.550038] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:61:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c964bdc6-fccc-40d9-bfe2-763b6f05a863', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69314b3d-cb79-40f5-8562-afed90be617f', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.557457] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Creating folder: Project (ef503fd02adc475fa79dda8b379b6514). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.558510] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ff71b0f6-5c82-4507-89a5-60c36b7a4c77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.570024] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293721, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503437} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.571346] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 413eaa2e-7bbc-402e-b0d3-f030b6395d7b/413eaa2e-7bbc-402e-b0d3-f030b6395d7b.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.571566] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.572825] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6300b42-9b15-4287-adf8-cf5dffa9d52b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.574641] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Created folder: Project (ef503fd02adc475fa79dda8b379b6514) in parent group-v845547. [ 835.574819] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Creating folder: Instances. Parent ref: group-v845688. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.575269] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-769508aa-49de-44ac-a80a-4b6c1e5cdaff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.582944] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for the task: (returnval){ [ 835.582944] env[62814]: value = "task-4293723" [ 835.582944] env[62814]: _type = "Task" [ 835.582944] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.587573] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Created folder: Instances in parent group-v845688. [ 835.587797] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 835.588591] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.588820] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6390b4be-10c0-4723-bb79-e1c625dec3d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.606628] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293723, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.610820] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.610820] env[62814]: value = "task-4293727" [ 835.610820] env[62814]: _type = "Task" [ 835.610820] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.618331] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293727, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.654278] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293720, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.841569] env[62814]: DEBUG nova.compute.manager [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 835.873499] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 835.873820] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.874038] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 835.874283] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.874464] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 835.874683] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 835.874928] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 835.875109] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 835.875283] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 835.875450] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 835.875726] env[62814]: DEBUG nova.virt.hardware [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 835.876841] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad212a35-a106-4578-aee0-fd7b56566317 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.885126] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc9f6da-84ba-4eca-987d-b0d9ce03a265 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.042517] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.221s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 836.044924] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.798s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 836.046555] env[62814]: INFO nova.compute.claims [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.070692] env[62814]: INFO nova.scheduler.client.report [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Deleted allocations for instance b289800f-49ce-49e7-b6bc-a3b4ec84b434 [ 836.094514] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293723, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098044} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.094766] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.095632] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5559d626-0b0a-4930-84b8-34886c56657b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.121099] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] 413eaa2e-7bbc-402e-b0d3-f030b6395d7b/413eaa2e-7bbc-402e-b0d3-f030b6395d7b.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.127348] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f9aa874-cbe7-4bbe-bee3-8962b5a943b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.150406] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293727, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.154873] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for the task: (returnval){ [ 836.154873] env[62814]: value = "task-4293729" [ 836.154873] env[62814]: _type = "Task" [ 836.154873] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.164026] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293720, 'name': CloneVM_Task} progress is 95%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.169144] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293729, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.376919] env[62814]: DEBUG nova.network.neutron [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Updated VIF entry in instance network info cache for port 69314b3d-cb79-40f5-8562-afed90be617f. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 836.376919] env[62814]: DEBUG nova.network.neutron [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Updating instance_info_cache with network_info: [{"id": "69314b3d-cb79-40f5-8562-afed90be617f", "address": "fa:16:3e:fa:61:e3", "network": {"id": "4ddbe620-bc9d-4b38-a414-7799c4e7060d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-334880125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef503fd02adc475fa79dda8b379b6514", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c964bdc6-fccc-40d9-bfe2-763b6f05a863", "external-id": "cl2-zone-376", "segmentation_id": 376, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69314b3d-cb", "ovs_interfaceid": "69314b3d-cb79-40f5-8562-afed90be617f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.509673] env[62814]: DEBUG nova.network.neutron [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Successfully updated port: fa04dc3a-1732-4479-80e7-1d3a547ea8c5 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.584452] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7b7cc8e4-963f-439c-9b35-91852bdf5458 tempest-ServerExternalEventsTest-1516125349 tempest-ServerExternalEventsTest-1516125349-project-member] Lock "b289800f-49ce-49e7-b6bc-a3b4ec84b434" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 41.555s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 836.623893] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293727, 'name': CreateVM_Task, 'duration_secs': 0.559616} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.624270] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 836.625151] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.625883] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 836.625883] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 836.626069] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89474279-c1d0-4efe-90e8-0cd323d361e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.631132] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 836.631132] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528d5b11-aa93-fedb-0e97-e8dac9b93763" [ 836.631132] env[62814]: _type = "Task" [ 836.631132] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.639182] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528d5b11-aa93-fedb-0e97-e8dac9b93763, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.658954] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293720, 'name': CloneVM_Task, 'duration_secs': 1.694049} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.661989] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Created linked-clone VM from snapshot [ 836.662814] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d2a260-4e28-499a-89b6-b05462228fec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.669715] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293729, 'name': ReconfigVM_Task, 'duration_secs': 0.29555} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.672906] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Reconfigured VM instance instance-00000032 to attach disk [datastore2] 413eaa2e-7bbc-402e-b0d3-f030b6395d7b/413eaa2e-7bbc-402e-b0d3-f030b6395d7b.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.673965] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Uploading image ac323e2a-0e96-4815-b9c0-bfde52e3c1c1 {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 836.675957] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f9a2ca7-79fd-4124-8c2c-9295ba3bd8b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.683911] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for the task: (returnval){ [ 836.683911] env[62814]: value = "task-4293730" [ 836.683911] env[62814]: _type = "Task" [ 836.683911] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.687687] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 836.688036] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2201e961-f93f-4d0b-832d-c4ef450510ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.695841] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293730, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.697039] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 836.697039] env[62814]: value = "task-4293731" [ 836.697039] env[62814]: _type = "Task" [ 836.697039] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.705351] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293731, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.877374] env[62814]: DEBUG oslo_concurrency.lockutils [req-99aad8dc-734e-4889-a00a-a109b0f1b6ac req-c531bb77-69ec-4b42-82f2-6059602b6af1 service nova] Releasing lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 837.012695] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "refresh_cache-c8765756-0870-4a06-a1a5-d02177959b29" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.012890] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "refresh_cache-c8765756-0870-4a06-a1a5-d02177959b29" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 837.013767] env[62814]: DEBUG nova.network.neutron [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.084583] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.084583] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.141849] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528d5b11-aa93-fedb-0e97-e8dac9b93763, 'name': SearchDatastore_Task, 'duration_secs': 0.010922} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.144471] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 837.144713] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.144940] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.145101] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 837.145286] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.146019] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7db28b23-dd54-4c8b-8799-1891e948fa07 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.155913] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.156186] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.156800] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-202eb174-3694-447a-8f77-5ba81abe9b36 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.164955] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 837.164955] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5237cc94-d674-a9ba-b7c0-35e1ec7947ce" [ 837.164955] env[62814]: _type = "Task" [ 837.164955] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.171942] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5237cc94-d674-a9ba-b7c0-35e1ec7947ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.194972] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293730, 'name': Rename_Task, 'duration_secs': 0.177065} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.197700] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.198195] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a0d0e76-fd44-4358-8959-08d691fa066a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.212023] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293731, 'name': Destroy_Task, 'duration_secs': 0.360094} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.212023] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Destroyed the VM [ 837.212023] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 837.212023] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for the task: (returnval){ [ 837.212023] env[62814]: value = "task-4293732" [ 837.212023] env[62814]: _type = "Task" [ 837.212023] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.213546] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4bb3f925-15c3-407f-a495-c186363248d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.224990] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293732, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.226340] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 837.226340] env[62814]: value = "task-4293733" [ 837.226340] env[62814]: _type = "Task" [ 837.226340] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.238414] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293733, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.504262] env[62814]: DEBUG nova.compute.manager [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Received event network-vif-plugged-fa04dc3a-1732-4479-80e7-1d3a547ea8c5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 837.504565] env[62814]: DEBUG oslo_concurrency.lockutils [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] Acquiring lock "c8765756-0870-4a06-a1a5-d02177959b29-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 837.504709] env[62814]: DEBUG oslo_concurrency.lockutils [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] Lock "c8765756-0870-4a06-a1a5-d02177959b29-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 837.505946] env[62814]: DEBUG oslo_concurrency.lockutils [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] Lock "c8765756-0870-4a06-a1a5-d02177959b29-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.505946] env[62814]: DEBUG nova.compute.manager [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] No waiting events found dispatching network-vif-plugged-fa04dc3a-1732-4479-80e7-1d3a547ea8c5 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 837.505946] env[62814]: WARNING nova.compute.manager [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Received unexpected event network-vif-plugged-fa04dc3a-1732-4479-80e7-1d3a547ea8c5 for instance with vm_state building and task_state spawning. [ 837.505946] env[62814]: DEBUG nova.compute.manager [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Received event network-changed-fa04dc3a-1732-4479-80e7-1d3a547ea8c5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 837.505946] env[62814]: DEBUG nova.compute.manager [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Refreshing instance network info cache due to event network-changed-fa04dc3a-1732-4479-80e7-1d3a547ea8c5. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 837.505946] env[62814]: DEBUG oslo_concurrency.lockutils [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] Acquiring lock "refresh_cache-c8765756-0870-4a06-a1a5-d02177959b29" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.550783] env[62814]: DEBUG nova.network.neutron [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.570471] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a710cf6-de91-4869-ba57-c4fff6b45f7a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.584365] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6977e5-0f49-4ff9-bad1-9334ed5c3571 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.592710] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.592885] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.593048] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.593842] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.593842] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.593842] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.593842] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 837.593842] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 837.629382] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0cb7bab-4f04-46e4-be89-4d4f7f7500a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.639314] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9ec08c-8034-48ea-9fd7-2275e3873bb5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.654424] env[62814]: DEBUG nova.compute.provider_tree [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 837.680334] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5237cc94-d674-a9ba-b7c0-35e1ec7947ce, 'name': SearchDatastore_Task, 'duration_secs': 0.009578} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.681333] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f63a5b6-30d5-42eb-8ccd-75d7a6337989 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.691174] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 837.691174] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52be59f5-5a7e-6da4-c4c0-1f88a91ee950" [ 837.691174] env[62814]: _type = "Task" [ 837.691174] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.701686] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52be59f5-5a7e-6da4-c4c0-1f88a91ee950, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.726756] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293732, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.736815] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293733, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.766868] env[62814]: DEBUG nova.network.neutron [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Updating instance_info_cache with network_info: [{"id": "fa04dc3a-1732-4479-80e7-1d3a547ea8c5", "address": "fa:16:3e:97:37:20", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa04dc3a-17", "ovs_interfaceid": "fa04dc3a-1732-4479-80e7-1d3a547ea8c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.126874] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.183725] env[62814]: ERROR nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [req-08b3f22c-54d9-4786-8e4f-cbed20af7d4f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-08b3f22c-54d9-4786-8e4f-cbed20af7d4f"}]} [ 838.207987] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52be59f5-5a7e-6da4-c4c0-1f88a91ee950, 'name': SearchDatastore_Task, 'duration_secs': 0.012536} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.211019] env[62814]: DEBUG nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 838.211756] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 838.212207] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 425e8edd-c002-45a1-bb6f-ee3ac8812509/425e8edd-c002-45a1-bb6f-ee3ac8812509.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 838.215017] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2218cfee-0de2-4a51-aee1-05df7bcc0218 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.222812] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 838.222812] env[62814]: value = "task-4293735" [ 838.222812] env[62814]: _type = "Task" [ 838.222812] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.230507] env[62814]: DEBUG nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 838.231073] env[62814]: DEBUG nova.compute.provider_tree [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 838.233804] env[62814]: DEBUG oslo_vmware.api [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293732, 'name': PowerOnVM_Task, 'duration_secs': 0.520604} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.238210] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.238415] env[62814]: INFO nova.compute.manager [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Took 7.61 seconds to spawn the instance on the hypervisor. [ 838.238856] env[62814]: DEBUG nova.compute.manager [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 838.242876] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063a7167-783a-4b0c-8cd4-53eca84abbb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.245715] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293735, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.247735] env[62814]: DEBUG nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 838.257329] env[62814]: DEBUG oslo_vmware.api [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293733, 'name': RemoveSnapshot_Task, 'duration_secs': 0.692105} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.257727] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 838.271140] env[62814]: DEBUG nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 838.273947] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "refresh_cache-c8765756-0870-4a06-a1a5-d02177959b29" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 838.275141] env[62814]: DEBUG nova.compute.manager [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Instance network_info: |[{"id": "fa04dc3a-1732-4479-80e7-1d3a547ea8c5", "address": "fa:16:3e:97:37:20", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa04dc3a-17", "ovs_interfaceid": "fa04dc3a-1732-4479-80e7-1d3a547ea8c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 838.275935] env[62814]: DEBUG oslo_concurrency.lockutils [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] Acquired lock "refresh_cache-c8765756-0870-4a06-a1a5-d02177959b29" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 838.276216] env[62814]: DEBUG nova.network.neutron [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Refreshing network info cache for port fa04dc3a-1732-4479-80e7-1d3a547ea8c5 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 838.277530] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:37:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3db2ab9e-1244-4377-b05f-ab76003f2428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fa04dc3a-1732-4479-80e7-1d3a547ea8c5', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 838.289095] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 838.289638] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 838.289882] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a8e80608-350a-4158-a796-172409735200 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.315427] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 838.315427] env[62814]: value = "task-4293736" [ 838.315427] env[62814]: _type = "Task" [ 838.315427] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.323516] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293736, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.737059] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293735, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.769947] env[62814]: WARNING nova.compute.manager [None req-c6b005df-2fbd-4084-a9ae-c0a397e8feb3 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Image not found during snapshot: nova.exception.ImageNotFound: Image ac323e2a-0e96-4815-b9c0-bfde52e3c1c1 could not be found. [ 838.772667] env[62814]: INFO nova.compute.manager [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Took 43.78 seconds to build instance. [ 838.826385] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293736, 'name': CreateVM_Task, 'duration_secs': 0.466441} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.829099] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.831928] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.832509] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 838.832509] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 838.832843] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fbb955aa-fcb9-40c7-8353-86c7d7a0613a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.837865] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 838.837865] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a2fbe7-0b77-f3ca-b2ed-ddb897cb0847" [ 838.837865] env[62814]: _type = "Task" [ 838.837865] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.848385] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a2fbe7-0b77-f3ca-b2ed-ddb897cb0847, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.933454] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9b314d-bb9b-4b0b-8009-605cb677c104 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.941304] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7a749a-af94-4efc-a1f5-f8840e081a07 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.981642] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b107a3-9cfa-4c7f-9ee8-e0e09fbac343 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.989580] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-935131e8-3639-4cee-b4b4-3454f29813b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.004211] env[62814]: DEBUG nova.compute.provider_tree [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 839.129112] env[62814]: DEBUG nova.network.neutron [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Updated VIF entry in instance network info cache for port fa04dc3a-1732-4479-80e7-1d3a547ea8c5. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 839.129470] env[62814]: DEBUG nova.network.neutron [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Updating instance_info_cache with network_info: [{"id": "fa04dc3a-1732-4479-80e7-1d3a547ea8c5", "address": "fa:16:3e:97:37:20", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfa04dc3a-17", "ovs_interfaceid": "fa04dc3a-1732-4479-80e7-1d3a547ea8c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.151492] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "98884571-8101-4ae9-b2e3-aeed97e3618f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 839.151733] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "98884571-8101-4ae9-b2e3-aeed97e3618f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 839.233775] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293735, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541807} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.234045] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 425e8edd-c002-45a1-bb6f-ee3ac8812509/425e8edd-c002-45a1-bb6f-ee3ac8812509.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.234256] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.234500] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c09732a6-c444-4641-9339-89789cf8dc9c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.240542] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 839.240542] env[62814]: value = "task-4293737" [ 839.240542] env[62814]: _type = "Task" [ 839.240542] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.247841] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293737, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.274165] env[62814]: DEBUG oslo_concurrency.lockutils [None req-efeca152-b896-45c0-a068-4db0dd239883 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.587s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.348297] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a2fbe7-0b77-f3ca-b2ed-ddb897cb0847, 'name': SearchDatastore_Task, 'duration_secs': 0.009614} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.348605] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.348836] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.349084] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.349514] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 839.349514] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.349925] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c318846-ea6c-494c-8e02-6b2299aab3e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.357673] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.357867] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.358560] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3177ddc5-f10f-4b2a-8500-ee917f67bb8b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.364549] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 839.364549] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5220d5ed-dd11-cec1-4714-58effdc763a1" [ 839.364549] env[62814]: _type = "Task" [ 839.364549] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.371691] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5220d5ed-dd11-cec1-4714-58effdc763a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.397313] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 839.397582] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 839.397828] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 839.398106] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 839.398395] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 839.400696] env[62814]: INFO nova.compute.manager [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Terminating instance [ 839.525831] env[62814]: ERROR nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [req-cfe68c9c-1020-4e2f-8a1b-9e89a08a5acf] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cfe68c9c-1020-4e2f-8a1b-9e89a08a5acf"}]} [ 839.542484] env[62814]: DEBUG nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 839.556780] env[62814]: DEBUG nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 839.557016] env[62814]: DEBUG nova.compute.provider_tree [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 839.570086] env[62814]: DEBUG nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 839.596351] env[62814]: DEBUG nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 839.632417] env[62814]: DEBUG oslo_concurrency.lockutils [req-6467a4bf-04ac-4c28-93d7-b18405f59ecf req-33ff245f-04cd-49b2-aaa9-80697cd26f92 service nova] Releasing lock "refresh_cache-c8765756-0870-4a06-a1a5-d02177959b29" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.752382] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293737, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.776496] env[62814]: DEBUG nova.compute.manager [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 839.877827] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5220d5ed-dd11-cec1-4714-58effdc763a1, 'name': SearchDatastore_Task, 'duration_secs': 0.008204} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.878415] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d175b9b-1c4e-4f48-93b2-13fd05bb66d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.884556] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 839.884556] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aced7e-c1f6-a178-9fd9-63a9499a7e8b" [ 839.884556] env[62814]: _type = "Task" [ 839.884556] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.898717] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aced7e-c1f6-a178-9fd9-63a9499a7e8b, 'name': SearchDatastore_Task, 'duration_secs': 0.01} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.898974] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.899247] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c8765756-0870-4a06-a1a5-d02177959b29/c8765756-0870-4a06-a1a5-d02177959b29.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.899499] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff393d09-b98b-4cc7-813a-7db52e34892f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.904735] env[62814]: DEBUG nova.compute.manager [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 839.904936] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 839.907217] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d57ec7a9-d653-45db-9acf-43efbbb89926 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.909881] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 839.909881] env[62814]: value = "task-4293738" [ 839.909881] env[62814]: _type = "Task" [ 839.909881] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.915055] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.915597] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ad3a362-737a-4177-9906-c180eb409203 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.922099] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293738, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.926370] env[62814]: DEBUG oslo_vmware.api [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 839.926370] env[62814]: value = "task-4293739" [ 839.926370] env[62814]: _type = "Task" [ 839.926370] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.938783] env[62814]: DEBUG oslo_vmware.api [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293739, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.986482] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 839.986755] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845693', 'volume_id': '996a4d77-b6db-42c8-93dc-c289397a5dd4', 'name': 'volume-996a4d77-b6db-42c8-93dc-c289397a5dd4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d3f32fc-276a-49be-b471-01a5d6fc5069', 'attached_at': '', 'detached_at': '', 'volume_id': '996a4d77-b6db-42c8-93dc-c289397a5dd4', 'serial': '996a4d77-b6db-42c8-93dc-c289397a5dd4'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 839.987674] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d939c6-2de0-41a0-aedf-28e9ba0c0cc0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.006371] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3de0c0-8ec6-4a1e-b4b0-eb82a6fefcd6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.032792] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Reconfiguring VM instance instance-0000000d to attach disk [datastore1] volume-996a4d77-b6db-42c8-93dc-c289397a5dd4/volume-996a4d77-b6db-42c8-93dc-c289397a5dd4.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.035743] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-512d0e56-1236-459e-8c44-c44fe599ce38 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.056606] env[62814]: DEBUG oslo_vmware.api [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Waiting for the task: (returnval){ [ 840.056606] env[62814]: value = "task-4293740" [ 840.056606] env[62814]: _type = "Task" [ 840.056606] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.069192] env[62814]: DEBUG oslo_vmware.api [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Task: {'id': task-4293740, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.102454] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-542a25d0-c843-4fa7-bc4d-47052adf071e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.111569] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bd7680-3b14-4931-8632-6505a46aa6e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.149114] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f3f529-a761-42d2-9a20-765884961ce4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.158577] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be8ed1f-82ed-4f75-94e4-5d657b2c962b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.174754] env[62814]: DEBUG nova.compute.provider_tree [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 840.228810] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquiring lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 840.229125] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 840.229393] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquiring lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 840.229602] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 840.229749] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 840.232216] env[62814]: INFO nova.compute.manager [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Terminating instance [ 840.253502] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293737, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.592109} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.253795] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.254622] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e47d94-8882-4f91-8476-e295bbf64794 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.279205] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 425e8edd-c002-45a1-bb6f-ee3ac8812509/425e8edd-c002-45a1-bb6f-ee3ac8812509.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.279552] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb545a13-5843-41c0-a031-e7d30ee36717 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.305372] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 840.305372] env[62814]: value = "task-4293741" [ 840.305372] env[62814]: _type = "Task" [ 840.305372] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.316149] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293741, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.317400] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 840.423536] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293738, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515507} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.424492] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c8765756-0870-4a06-a1a5-d02177959b29/c8765756-0870-4a06-a1a5-d02177959b29.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.424492] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.424727] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43995b40-7a5c-4cd4-8d3e-9d81a0332847 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.435415] env[62814]: DEBUG oslo_vmware.api [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293739, 'name': PowerOffVM_Task, 'duration_secs': 0.286179} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.437117] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 840.437325] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 840.437654] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 840.437654] env[62814]: value = "task-4293742" [ 840.437654] env[62814]: _type = "Task" [ 840.437654] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.437860] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16835cd1-f3e9-44bd-afdd-b6556b9ef115 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.447919] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293742, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.514870] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 840.515301] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 840.515572] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Deleting the datastore file [datastore2] 7cca1c35-6bfc-450d-ba74-0e825b160e8f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.515927] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cef8528f-a120-4ca7-9c5f-76a0bf8f407f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.523045] env[62814]: DEBUG oslo_vmware.api [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 840.523045] env[62814]: value = "task-4293744" [ 840.523045] env[62814]: _type = "Task" [ 840.523045] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.534143] env[62814]: DEBUG oslo_vmware.api [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293744, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.566253] env[62814]: DEBUG oslo_vmware.api [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Task: {'id': task-4293740, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.715116] env[62814]: DEBUG nova.scheduler.client.report [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 82 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 840.715414] env[62814]: DEBUG nova.compute.provider_tree [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 82 to 83 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 840.716027] env[62814]: DEBUG nova.compute.provider_tree [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 840.736796] env[62814]: DEBUG nova.compute.manager [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 840.737023] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.738248] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e93171-fe97-47ec-9d3e-23417198f5a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.746464] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.746708] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10f0f85b-cb6c-41de-b328-092f143868fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.753424] env[62814]: DEBUG oslo_vmware.api [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for the task: (returnval){ [ 840.753424] env[62814]: value = "task-4293745" [ 840.753424] env[62814]: _type = "Task" [ 840.753424] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.762920] env[62814]: DEBUG oslo_vmware.api [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293745, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.815930] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293741, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.948835] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293742, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074192} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.949199] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.950074] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a0a34dd-5ef2-44e1-bf56-d01555e27afb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.973440] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] c8765756-0870-4a06-a1a5-d02177959b29/c8765756-0870-4a06-a1a5-d02177959b29.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.973763] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-679ba36d-4f68-4135-9c5b-84bcd39d73a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.992796] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 840.992796] env[62814]: value = "task-4293746" [ 840.992796] env[62814]: _type = "Task" [ 840.992796] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.002254] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293746, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.032513] env[62814]: DEBUG oslo_vmware.api [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293744, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186027} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.032513] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.032733] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.032968] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.033065] env[62814]: INFO nova.compute.manager [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 841.034023] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 841.034023] env[62814]: DEBUG nova.compute.manager [-] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 841.034023] env[62814]: DEBUG nova.network.neutron [-] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.070172] env[62814]: DEBUG oslo_vmware.api [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Task: {'id': task-4293740, 'name': ReconfigVM_Task, 'duration_secs': 0.67599} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.070543] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Reconfigured VM instance instance-0000000d to attach disk [datastore1] volume-996a4d77-b6db-42c8-93dc-c289397a5dd4/volume-996a4d77-b6db-42c8-93dc-c289397a5dd4.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.075401] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9eab25e8-5194-42a2-b3ca-3461b0475810 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.090822] env[62814]: DEBUG oslo_vmware.api [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Waiting for the task: (returnval){ [ 841.090822] env[62814]: value = "task-4293747" [ 841.090822] env[62814]: _type = "Task" [ 841.090822] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.098821] env[62814]: DEBUG oslo_vmware.api [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Task: {'id': task-4293747, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.220994] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 5.176s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.221576] env[62814]: DEBUG nova.compute.manager [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 841.225925] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.734s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.227365] env[62814]: INFO nova.compute.claims [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 841.264282] env[62814]: DEBUG oslo_vmware.api [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293745, 'name': PowerOffVM_Task, 'duration_secs': 0.216995} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.264834] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 841.265026] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 841.265236] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64525205-cef0-4124-ae54-f915f1a021f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.316367] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293741, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.330013] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.330013] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.330013] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Deleting the datastore file [datastore2] 413eaa2e-7bbc-402e-b0d3-f030b6395d7b {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.330013] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-044b8d31-6911-418f-b633-bce1cfd303b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.335724] env[62814]: DEBUG oslo_vmware.api [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for the task: (returnval){ [ 841.335724] env[62814]: value = "task-4293749" [ 841.335724] env[62814]: _type = "Task" [ 841.335724] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.347277] env[62814]: DEBUG oslo_vmware.api [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293749, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.355588] env[62814]: DEBUG nova.compute.manager [req-7db03e20-9c9b-4b31-a3df-4295d4fc0209 req-caafeaed-a0cf-486e-8d7f-b502800ff54b service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Received event network-vif-deleted-87aa2371-717e-4ae6-9aab-1b22613e710c {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 841.356473] env[62814]: INFO nova.compute.manager [req-7db03e20-9c9b-4b31-a3df-4295d4fc0209 req-caafeaed-a0cf-486e-8d7f-b502800ff54b service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Neutron deleted interface 87aa2371-717e-4ae6-9aab-1b22613e710c; detaching it from the instance and deleting it from the info cache [ 841.356473] env[62814]: DEBUG nova.network.neutron [req-7db03e20-9c9b-4b31-a3df-4295d4fc0209 req-caafeaed-a0cf-486e-8d7f-b502800ff54b service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.503955] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293746, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.600355] env[62814]: DEBUG oslo_vmware.api [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Task: {'id': task-4293747, 'name': ReconfigVM_Task, 'duration_secs': 0.140238} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.600702] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845693', 'volume_id': '996a4d77-b6db-42c8-93dc-c289397a5dd4', 'name': 'volume-996a4d77-b6db-42c8-93dc-c289397a5dd4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d3f32fc-276a-49be-b471-01a5d6fc5069', 'attached_at': '', 'detached_at': '', 'volume_id': '996a4d77-b6db-42c8-93dc-c289397a5dd4', 'serial': '996a4d77-b6db-42c8-93dc-c289397a5dd4'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 841.732969] env[62814]: DEBUG nova.compute.utils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 841.737788] env[62814]: DEBUG nova.compute.manager [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 841.737788] env[62814]: DEBUG nova.network.neutron [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 841.792176] env[62814]: DEBUG nova.policy [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8497da37b9de4640826067a536d8a358', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b3ba1c9c7624abfb3e0bd880dc40166', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 841.818256] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293741, 'name': ReconfigVM_Task, 'duration_secs': 1.366982} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.818395] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 425e8edd-c002-45a1-bb6f-ee3ac8812509/425e8edd-c002-45a1-bb6f-ee3ac8812509.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.818966] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f192b664-c448-479e-8007-b4f02181193e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.825921] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 841.825921] env[62814]: value = "task-4293750" [ 841.825921] env[62814]: _type = "Task" [ 841.825921] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.834703] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293750, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.845490] env[62814]: DEBUG oslo_vmware.api [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Task: {'id': task-4293749, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139761} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.845748] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.845991] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.846204] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.846435] env[62814]: INFO nova.compute.manager [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 841.846736] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 841.846963] env[62814]: DEBUG nova.compute.manager [-] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 841.847098] env[62814]: DEBUG nova.network.neutron [-] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.849036] env[62814]: DEBUG nova.network.neutron [-] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.859363] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b347cd34-0a10-4806-baf0-196f8b1a08a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.870255] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a985ed4-da8d-44de-88bf-90647cb36797 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.908971] env[62814]: DEBUG nova.compute.manager [req-7db03e20-9c9b-4b31-a3df-4295d4fc0209 req-caafeaed-a0cf-486e-8d7f-b502800ff54b service nova] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Detach interface failed, port_id=87aa2371-717e-4ae6-9aab-1b22613e710c, reason: Instance 7cca1c35-6bfc-450d-ba74-0e825b160e8f could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 842.004843] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293746, 'name': ReconfigVM_Task, 'duration_secs': 0.864428} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.005134] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Reconfigured VM instance instance-00000034 to attach disk [datastore2] c8765756-0870-4a06-a1a5-d02177959b29/c8765756-0870-4a06-a1a5-d02177959b29.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.005761] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4363d661-8916-46a7-a0eb-93b5da034a82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.012886] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 842.012886] env[62814]: value = "task-4293751" [ 842.012886] env[62814]: _type = "Task" [ 842.012886] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.021360] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293751, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.229249] env[62814]: DEBUG nova.compute.manager [req-9be99141-61ff-4533-a039-3316d9b01485 req-cf162301-cabc-4d9d-9614-98be6a708a82 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Received event network-vif-deleted-a5c9c9ce-5548-4870-b39b-8d82d5a6bff4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 842.229387] env[62814]: INFO nova.compute.manager [req-9be99141-61ff-4533-a039-3316d9b01485 req-cf162301-cabc-4d9d-9614-98be6a708a82 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Neutron deleted interface a5c9c9ce-5548-4870-b39b-8d82d5a6bff4; detaching it from the instance and deleting it from the info cache [ 842.229593] env[62814]: DEBUG nova.network.neutron [req-9be99141-61ff-4533-a039-3316d9b01485 req-cf162301-cabc-4d9d-9614-98be6a708a82 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.238429] env[62814]: DEBUG nova.compute.manager [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 842.242604] env[62814]: DEBUG nova.network.neutron [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Successfully created port: 9fed46ec-04e6-4482-a460-370e4ff04a32 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.335205] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293750, 'name': Rename_Task, 'duration_secs': 0.136529} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.335495] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.337979] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc272c74-87da-4c63-a1ac-af3f0cb0428e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.345149] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 842.345149] env[62814]: value = "task-4293752" [ 842.345149] env[62814]: _type = "Task" [ 842.345149] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.352463] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293752, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.357158] env[62814]: INFO nova.compute.manager [-] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Took 1.32 seconds to deallocate network for instance. [ 842.525505] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293751, 'name': Rename_Task, 'duration_secs': 0.279251} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.526355] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.526355] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-20bdac51-61fe-4b05-98f6-5c29e14111b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.533062] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 842.533062] env[62814]: value = "task-4293753" [ 842.533062] env[62814]: _type = "Task" [ 842.533062] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.544203] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.649684] env[62814]: DEBUG nova.network.neutron [-] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.655520] env[62814]: DEBUG nova.objects.instance [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Lazy-loading 'flavor' on Instance uuid 3d3f32fc-276a-49be-b471-01a5d6fc5069 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.731973] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f105914-f27f-4bd6-bc3d-a03e1e7d8c12 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.744549] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfb5e50-2dd1-4771-9dda-6116c87d3f95 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.770664] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4709c4f6-8105-4391-8e64-8dbb86007fcb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.786811] env[62814]: DEBUG nova.compute.manager [req-9be99141-61ff-4533-a039-3316d9b01485 req-cf162301-cabc-4d9d-9614-98be6a708a82 service nova] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Detach interface failed, port_id=a5c9c9ce-5548-4870-b39b-8d82d5a6bff4, reason: Instance 413eaa2e-7bbc-402e-b0d3-f030b6395d7b could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 842.791936] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b4b0646-bb63-40cf-a632-5330b0d55084 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.824491] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe75ab94-d71a-478f-b6ab-c637638746f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.832655] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e6ff6e-412f-45f6-9518-9220b8c2ca79 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.846407] env[62814]: DEBUG nova.compute.provider_tree [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.856865] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293752, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.862729] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.043458] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.157032] env[62814]: INFO nova.compute.manager [-] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Took 1.31 seconds to deallocate network for instance. [ 843.161474] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c7dcef0f-4314-460d-b868-0e3cd3c92290 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.808s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.221471] env[62814]: DEBUG oslo_concurrency.lockutils [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Acquiring lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.221808] env[62814]: DEBUG oslo_concurrency.lockutils [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.263358] env[62814]: DEBUG nova.compute.manager [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 843.293064] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 843.293346] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 843.293503] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 843.293685] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 843.293824] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 843.293970] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 843.294208] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 843.294370] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 843.294538] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 843.294701] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 843.294875] env[62814]: DEBUG nova.virt.hardware [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 843.295775] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7a2f9c-59d5-42ea-9346-f8e3cf54fdc9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.303986] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0514c089-7320-4a2e-8d69-82bfc0529bbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.352399] env[62814]: DEBUG nova.scheduler.client.report [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 843.358524] env[62814]: DEBUG oslo_vmware.api [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293752, 'name': PowerOnVM_Task, 'duration_secs': 0.750128} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.358762] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.358957] env[62814]: INFO nova.compute.manager [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Took 10.06 seconds to spawn the instance on the hypervisor. [ 843.359592] env[62814]: DEBUG nova.compute.manager [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 843.360593] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c947e66f-650d-4b06-b919-de6260043345 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.546016] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293753, 'name': PowerOnVM_Task} progress is 74%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.678487] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.719687] env[62814]: DEBUG nova.compute.manager [req-d28cc6b8-5446-4c82-935c-1bb397f25fa3 req-d189c654-3438-48a4-af08-4a443ec8512c service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Received event network-vif-plugged-9fed46ec-04e6-4482-a460-370e4ff04a32 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 843.719876] env[62814]: DEBUG oslo_concurrency.lockutils [req-d28cc6b8-5446-4c82-935c-1bb397f25fa3 req-d189c654-3438-48a4-af08-4a443ec8512c service nova] Acquiring lock "9683894b-a300-4400-a1b9-db62478f42c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 843.720142] env[62814]: DEBUG oslo_concurrency.lockutils [req-d28cc6b8-5446-4c82-935c-1bb397f25fa3 req-d189c654-3438-48a4-af08-4a443ec8512c service nova] Lock "9683894b-a300-4400-a1b9-db62478f42c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.720331] env[62814]: DEBUG oslo_concurrency.lockutils [req-d28cc6b8-5446-4c82-935c-1bb397f25fa3 req-d189c654-3438-48a4-af08-4a443ec8512c service nova] Lock "9683894b-a300-4400-a1b9-db62478f42c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.720545] env[62814]: DEBUG nova.compute.manager [req-d28cc6b8-5446-4c82-935c-1bb397f25fa3 req-d189c654-3438-48a4-af08-4a443ec8512c service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] No waiting events found dispatching network-vif-plugged-9fed46ec-04e6-4482-a460-370e4ff04a32 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 843.720774] env[62814]: WARNING nova.compute.manager [req-d28cc6b8-5446-4c82-935c-1bb397f25fa3 req-d189c654-3438-48a4-af08-4a443ec8512c service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Received unexpected event network-vif-plugged-9fed46ec-04e6-4482-a460-370e4ff04a32 for instance with vm_state building and task_state spawning. [ 843.725470] env[62814]: INFO nova.compute.manager [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Detaching volume 996a4d77-b6db-42c8-93dc-c289397a5dd4 [ 843.770491] env[62814]: INFO nova.virt.block_device [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Attempting to driver detach volume 996a4d77-b6db-42c8-93dc-c289397a5dd4 from mountpoint /dev/sdb [ 843.770692] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 843.770884] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845693', 'volume_id': '996a4d77-b6db-42c8-93dc-c289397a5dd4', 'name': 'volume-996a4d77-b6db-42c8-93dc-c289397a5dd4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d3f32fc-276a-49be-b471-01a5d6fc5069', 'attached_at': '', 'detached_at': '', 'volume_id': '996a4d77-b6db-42c8-93dc-c289397a5dd4', 'serial': '996a4d77-b6db-42c8-93dc-c289397a5dd4'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 843.771770] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e12a94-771d-405c-a281-b38ef9dd3562 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.793904] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58932fce-0335-46e6-a49d-7df2d10826ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.801101] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6682fa0b-4285-4c91-bf89-14b49b87eb1a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.821117] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14832d4-e27a-4828-a91f-d69829599dc8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.824356] env[62814]: DEBUG nova.network.neutron [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Successfully updated port: 9fed46ec-04e6-4482-a460-370e4ff04a32 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.837684] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] The volume has not been displaced from its original location: [datastore1] volume-996a4d77-b6db-42c8-93dc-c289397a5dd4/volume-996a4d77-b6db-42c8-93dc-c289397a5dd4.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 843.843393] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Reconfiguring VM instance instance-0000000d to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 843.844376] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cf3240c-3deb-43e1-a0e2-fdf3cf5b33ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.861181] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.635s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.862617] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 37.499s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 843.862804] env[62814]: DEBUG nova.objects.instance [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 843.866826] env[62814]: DEBUG oslo_vmware.api [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Waiting for the task: (returnval){ [ 843.866826] env[62814]: value = "task-4293754" [ 843.866826] env[62814]: _type = "Task" [ 843.866826] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.879235] env[62814]: DEBUG oslo_vmware.api [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Task: {'id': task-4293754, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.879936] env[62814]: INFO nova.compute.manager [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Took 48.31 seconds to build instance. [ 844.044792] env[62814]: DEBUG oslo_vmware.api [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293753, 'name': PowerOnVM_Task, 'duration_secs': 1.198059} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.045153] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 844.045403] env[62814]: INFO nova.compute.manager [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Took 8.20 seconds to spawn the instance on the hypervisor. [ 844.045624] env[62814]: DEBUG nova.compute.manager [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 844.046735] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2edfda-2fce-4fbc-b677-4d60832a346b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.327122] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.327308] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquired lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 844.327482] env[62814]: DEBUG nova.network.neutron [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.368524] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquiring lock "dc065c7d-2571-40a7-b5c0-ff8c544a4fef" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 844.368524] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "dc065c7d-2571-40a7-b5c0-ff8c544a4fef" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.381849] env[62814]: DEBUG oslo_vmware.api [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Task: {'id': task-4293754, 'name': ReconfigVM_Task, 'duration_secs': 0.227575} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.382254] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7123d3db-2ca5-4f6e-ab20-15fcb20c8acc tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "425e8edd-c002-45a1-bb6f-ee3ac8812509" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 71.596s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.382511] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Reconfigured VM instance instance-0000000d to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 844.388345] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a36eccc-6f9a-44da-9dbf-d309e984de9c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.407232] env[62814]: DEBUG oslo_vmware.api [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Waiting for the task: (returnval){ [ 844.407232] env[62814]: value = "task-4293755" [ 844.407232] env[62814]: _type = "Task" [ 844.407232] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.415377] env[62814]: DEBUG oslo_vmware.api [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Task: {'id': task-4293755, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.564339] env[62814]: INFO nova.compute.manager [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Took 47.91 seconds to build instance. [ 844.786579] env[62814]: DEBUG nova.compute.manager [req-a73c37cf-96fe-4f64-9ffc-563d7b1243e7 req-d966cfca-e7a5-4d3b-9d41-57b959db61ea service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Received event network-changed-69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 844.786789] env[62814]: DEBUG nova.compute.manager [req-a73c37cf-96fe-4f64-9ffc-563d7b1243e7 req-d966cfca-e7a5-4d3b-9d41-57b959db61ea service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Refreshing instance network info cache due to event network-changed-69314b3d-cb79-40f5-8562-afed90be617f. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 844.787017] env[62814]: DEBUG oslo_concurrency.lockutils [req-a73c37cf-96fe-4f64-9ffc-563d7b1243e7 req-d966cfca-e7a5-4d3b-9d41-57b959db61ea service nova] Acquiring lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.787174] env[62814]: DEBUG oslo_concurrency.lockutils [req-a73c37cf-96fe-4f64-9ffc-563d7b1243e7 req-d966cfca-e7a5-4d3b-9d41-57b959db61ea service nova] Acquired lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 844.787334] env[62814]: DEBUG nova.network.neutron [req-a73c37cf-96fe-4f64-9ffc-563d7b1243e7 req-d966cfca-e7a5-4d3b-9d41-57b959db61ea service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Refreshing network info cache for port 69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.873066] env[62814]: DEBUG nova.network.neutron [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.875415] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "dc065c7d-2571-40a7-b5c0-ff8c544a4fef" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.507s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.875918] env[62814]: DEBUG nova.compute.manager [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 844.879250] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a03f244d-fd4c-43e5-aa70-1bcb4b72e98e tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.881357] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 35.189s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.882822] env[62814]: INFO nova.compute.claims [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.900849] env[62814]: DEBUG nova.compute.manager [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 844.917706] env[62814]: DEBUG oslo_vmware.api [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Task: {'id': task-4293755, 'name': ReconfigVM_Task, 'duration_secs': 0.197494} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.918069] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845693', 'volume_id': '996a4d77-b6db-42c8-93dc-c289397a5dd4', 'name': 'volume-996a4d77-b6db-42c8-93dc-c289397a5dd4', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3d3f32fc-276a-49be-b471-01a5d6fc5069', 'attached_at': '', 'detached_at': '', 'volume_id': '996a4d77-b6db-42c8-93dc-c289397a5dd4', 'serial': '996a4d77-b6db-42c8-93dc-c289397a5dd4'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 845.066061] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd9604ee-0794-4f39-97a3-93c4a3aa7d01 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "c8765756-0870-4a06-a1a5-d02177959b29" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.238s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.125099] env[62814]: DEBUG nova.network.neutron [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Updating instance_info_cache with network_info: [{"id": "9fed46ec-04e6-4482-a460-370e4ff04a32", "address": "fa:16:3e:19:a2:ac", "network": {"id": "5363651c-6ade-4a4c-b762-9fc7cbbcf5ea", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1833251571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b3ba1c9c7624abfb3e0bd880dc40166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fed46ec-04", "ovs_interfaceid": "9fed46ec-04e6-4482-a460-370e4ff04a32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.384202] env[62814]: DEBUG nova.compute.utils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 845.386032] env[62814]: DEBUG nova.compute.manager [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 845.386133] env[62814]: DEBUG nova.network.neutron [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 845.426701] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 845.434807] env[62814]: DEBUG nova.policy [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae3932e2c8cb4251888c77499fb09d7f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f39e30833c0d4ed8bd3e3870ab9084d1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 845.526797] env[62814]: DEBUG nova.objects.instance [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Lazy-loading 'flavor' on Instance uuid 3d3f32fc-276a-49be-b471-01a5d6fc5069 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.591315] env[62814]: DEBUG nova.network.neutron [req-a73c37cf-96fe-4f64-9ffc-563d7b1243e7 req-d966cfca-e7a5-4d3b-9d41-57b959db61ea service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Updated VIF entry in instance network info cache for port 69314b3d-cb79-40f5-8562-afed90be617f. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.591777] env[62814]: DEBUG nova.network.neutron [req-a73c37cf-96fe-4f64-9ffc-563d7b1243e7 req-d966cfca-e7a5-4d3b-9d41-57b959db61ea service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Updating instance_info_cache with network_info: [{"id": "69314b3d-cb79-40f5-8562-afed90be617f", "address": "fa:16:3e:fa:61:e3", "network": {"id": "4ddbe620-bc9d-4b38-a414-7799c4e7060d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-334880125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef503fd02adc475fa79dda8b379b6514", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c964bdc6-fccc-40d9-bfe2-763b6f05a863", "external-id": "cl2-zone-376", "segmentation_id": 376, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69314b3d-cb", "ovs_interfaceid": "69314b3d-cb79-40f5-8562-afed90be617f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.627826] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Releasing lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 845.628183] env[62814]: DEBUG nova.compute.manager [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Instance network_info: |[{"id": "9fed46ec-04e6-4482-a460-370e4ff04a32", "address": "fa:16:3e:19:a2:ac", "network": {"id": "5363651c-6ade-4a4c-b762-9fc7cbbcf5ea", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1833251571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b3ba1c9c7624abfb3e0bd880dc40166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fed46ec-04", "ovs_interfaceid": "9fed46ec-04e6-4482-a460-370e4ff04a32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 845.628783] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:19:a2:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50cf0a70-948d-4611-af05-94c1483064ed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9fed46ec-04e6-4482-a460-370e4ff04a32', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.636034] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Creating folder: Project (2b3ba1c9c7624abfb3e0bd880dc40166). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.636633] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5aee2aa1-5e47-4668-9259-b813e1852042 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.648515] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Created folder: Project (2b3ba1c9c7624abfb3e0bd880dc40166) in parent group-v845547. [ 845.648715] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Creating folder: Instances. Parent ref: group-v845695. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.648955] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a3f4982a-a844-44c0-bb58-9bd6b27886a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.657650] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Created folder: Instances in parent group-v845695. [ 845.657892] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 845.658117] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.658330] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d239223-dbe4-430f-ac10-6e645cb8ae1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.679355] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.679355] env[62814]: value = "task-4293758" [ 845.679355] env[62814]: _type = "Task" [ 845.679355] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.690145] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293758, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.774066] env[62814]: DEBUG nova.network.neutron [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Successfully created port: b6e71b20-cd42-432e-bc6d-1c16b443ed92 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 845.893397] env[62814]: DEBUG nova.compute.manager [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 845.974227] env[62814]: DEBUG nova.compute.manager [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Received event network-changed-9fed46ec-04e6-4482-a460-370e4ff04a32 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 845.974227] env[62814]: DEBUG nova.compute.manager [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Refreshing instance network info cache due to event network-changed-9fed46ec-04e6-4482-a460-370e4ff04a32. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 845.974449] env[62814]: DEBUG oslo_concurrency.lockutils [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] Acquiring lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.974598] env[62814]: DEBUG oslo_concurrency.lockutils [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] Acquired lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 845.974799] env[62814]: DEBUG nova.network.neutron [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Refreshing network info cache for port 9fed46ec-04e6-4482-a460-370e4ff04a32 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.094673] env[62814]: DEBUG oslo_concurrency.lockutils [req-a73c37cf-96fe-4f64-9ffc-563d7b1243e7 req-d966cfca-e7a5-4d3b-9d41-57b959db61ea service nova] Releasing lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.191730] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293758, 'name': CreateVM_Task, 'duration_secs': 0.343921} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.192040] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.192624] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.193014] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.193220] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 846.193396] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf5a4e5a-3821-481c-9bc0-1632af489f85 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.198665] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 846.198665] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5221a230-bab0-4cb5-26b5-42229f125b04" [ 846.198665] env[62814]: _type = "Task" [ 846.198665] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.212291] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5221a230-bab0-4cb5-26b5-42229f125b04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.415135] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33144955-019c-4de1-9334-6b5b94a8f3cc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.422896] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e1916a3-7644-4b83-9f8f-23170eaa210b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.458099] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f24918-9bf2-4a96-8ff3-a6de12498130 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.466386] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6cd2e49-849a-439b-bae3-2648e5388d8e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.482978] env[62814]: DEBUG nova.compute.provider_tree [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.488022] env[62814]: DEBUG nova.compute.manager [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 846.488022] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1dec1b-c683-4dca-8aac-0519b2658ec0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.534922] env[62814]: DEBUG oslo_concurrency.lockutils [None req-224050ee-a52d-485a-b1f6-117f8b2cefd4 tempest-VolumesAssistedSnapshotsTest-1313881415 tempest-VolumesAssistedSnapshotsTest-1313881415-project-admin] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.313s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 846.712223] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5221a230-bab0-4cb5-26b5-42229f125b04, 'name': SearchDatastore_Task, 'duration_secs': 0.011608} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.712549] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 846.712785] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.713030] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.713182] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.713367] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.713627] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca636dd7-be93-4386-9083-4dac2cc2a878 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.724894] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.725099] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.725837] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e630a2a8-1322-400b-9f6f-4aac7b973abd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.731909] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 846.731909] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e9cf13-d42e-f382-49c1-72b2cbb7bdb9" [ 846.731909] env[62814]: _type = "Task" [ 846.731909] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.740366] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e9cf13-d42e-f382-49c1-72b2cbb7bdb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.767479] env[62814]: DEBUG nova.network.neutron [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Updated VIF entry in instance network info cache for port 9fed46ec-04e6-4482-a460-370e4ff04a32. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 846.767479] env[62814]: DEBUG nova.network.neutron [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Updating instance_info_cache with network_info: [{"id": "9fed46ec-04e6-4482-a460-370e4ff04a32", "address": "fa:16:3e:19:a2:ac", "network": {"id": "5363651c-6ade-4a4c-b762-9fc7cbbcf5ea", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1833251571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b3ba1c9c7624abfb3e0bd880dc40166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fed46ec-04", "ovs_interfaceid": "9fed46ec-04e6-4482-a460-370e4ff04a32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.907588] env[62814]: DEBUG nova.compute.manager [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 846.937229] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 846.939085] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 846.939085] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 846.939085] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 846.939085] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 846.939085] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 846.939085] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 846.939085] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 846.939085] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 846.939640] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 846.939640] env[62814]: DEBUG nova.virt.hardware [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 846.940308] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bffc60c7-bc7b-49f2-b320-a067c6253a3d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.948973] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd091a7-dbb9-4e88-ab0a-ad6486b98ada {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.989872] env[62814]: DEBUG nova.scheduler.client.report [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 847.002024] env[62814]: INFO nova.compute.manager [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] instance snapshotting [ 847.002024] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59bcb6b2-421e-461d-a5de-850f258844ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.022865] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0c07a13-71bb-492f-b364-f990bdf73c88 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.242275] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e9cf13-d42e-f382-49c1-72b2cbb7bdb9, 'name': SearchDatastore_Task, 'duration_secs': 0.009947} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.243099] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4839cee8-5319-4383-9555-1b804ae06681 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.248281] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 847.248281] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520ed176-f6db-4c21-7468-3d334535be9f" [ 847.248281] env[62814]: _type = "Task" [ 847.248281] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.255497] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520ed176-f6db-4c21-7468-3d334535be9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.269163] env[62814]: DEBUG oslo_concurrency.lockutils [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] Releasing lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 847.269481] env[62814]: DEBUG nova.compute.manager [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Received event network-changed-69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 847.269673] env[62814]: DEBUG nova.compute.manager [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Refreshing instance network info cache due to event network-changed-69314b3d-cb79-40f5-8562-afed90be617f. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 847.269885] env[62814]: DEBUG oslo_concurrency.lockutils [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] Acquiring lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.270039] env[62814]: DEBUG oslo_concurrency.lockutils [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] Acquired lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 847.270266] env[62814]: DEBUG nova.network.neutron [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Refreshing network info cache for port 69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.458728] env[62814]: DEBUG nova.compute.manager [req-7cc2391d-4fac-40cb-b041-37962d1e30d0 req-720e2aec-5ab2-4c83-bada-0a655a93e7e9 service nova] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Received event network-vif-plugged-b6e71b20-cd42-432e-bc6d-1c16b443ed92 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 847.459056] env[62814]: DEBUG oslo_concurrency.lockutils [req-7cc2391d-4fac-40cb-b041-37962d1e30d0 req-720e2aec-5ab2-4c83-bada-0a655a93e7e9 service nova] Acquiring lock "448b7bfb-1a90-4c5b-9086-9320cedaf015-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 847.459272] env[62814]: DEBUG oslo_concurrency.lockutils [req-7cc2391d-4fac-40cb-b041-37962d1e30d0 req-720e2aec-5ab2-4c83-bada-0a655a93e7e9 service nova] Lock "448b7bfb-1a90-4c5b-9086-9320cedaf015-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 847.459534] env[62814]: DEBUG oslo_concurrency.lockutils [req-7cc2391d-4fac-40cb-b041-37962d1e30d0 req-720e2aec-5ab2-4c83-bada-0a655a93e7e9 service nova] Lock "448b7bfb-1a90-4c5b-9086-9320cedaf015-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 847.459725] env[62814]: DEBUG nova.compute.manager [req-7cc2391d-4fac-40cb-b041-37962d1e30d0 req-720e2aec-5ab2-4c83-bada-0a655a93e7e9 service nova] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] No waiting events found dispatching network-vif-plugged-b6e71b20-cd42-432e-bc6d-1c16b443ed92 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 847.459893] env[62814]: WARNING nova.compute.manager [req-7cc2391d-4fac-40cb-b041-37962d1e30d0 req-720e2aec-5ab2-4c83-bada-0a655a93e7e9 service nova] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Received unexpected event network-vif-plugged-b6e71b20-cd42-432e-bc6d-1c16b443ed92 for instance with vm_state building and task_state spawning. [ 847.501556] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 847.502325] env[62814]: DEBUG nova.compute.manager [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 847.504546] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 35.416s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 847.505621] env[62814]: DEBUG nova.objects.instance [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lazy-loading 'resources' on Instance uuid 127fee64-fd56-4a23-bdd2-18c817898fd5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.537616] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 847.537897] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7e1f4bdb-bd3e-48ef-889e-76bd9c86ae6f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.547410] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 847.547410] env[62814]: value = "task-4293760" [ 847.547410] env[62814]: _type = "Task" [ 847.547410] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.557200] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293760, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.558411] env[62814]: DEBUG nova.network.neutron [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Successfully updated port: b6e71b20-cd42-432e-bc6d-1c16b443ed92 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.758941] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520ed176-f6db-4c21-7468-3d334535be9f, 'name': SearchDatastore_Task, 'duration_secs': 0.009502} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.759262] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 847.759540] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 9683894b-a300-4400-a1b9-db62478f42c5/9683894b-a300-4400-a1b9-db62478f42c5.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 847.760066] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dae36d81-1468-4408-b3d2-35206b750c3e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.766324] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 847.766324] env[62814]: value = "task-4293761" [ 847.766324] env[62814]: _type = "Task" [ 847.766324] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.777291] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293761, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.011137] env[62814]: DEBUG nova.compute.utils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 848.017893] env[62814]: DEBUG nova.compute.manager [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 848.021021] env[62814]: DEBUG nova.network.neutron [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.063366] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquiring lock "refresh_cache-448b7bfb-1a90-4c5b-9086-9320cedaf015" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.064050] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquired lock "refresh_cache-448b7bfb-1a90-4c5b-9086-9320cedaf015" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 848.064282] env[62814]: DEBUG nova.network.neutron [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.067480] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293760, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.087998] env[62814]: DEBUG nova.policy [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '713f3b7e15404bffa24349dac34c8534', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f11d1ef620764fa4b1e2b718ac207a44', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 848.114542] env[62814]: DEBUG nova.network.neutron [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Updated VIF entry in instance network info cache for port 69314b3d-cb79-40f5-8562-afed90be617f. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 848.114893] env[62814]: DEBUG nova.network.neutron [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Updating instance_info_cache with network_info: [{"id": "69314b3d-cb79-40f5-8562-afed90be617f", "address": "fa:16:3e:fa:61:e3", "network": {"id": "4ddbe620-bc9d-4b38-a414-7799c4e7060d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-334880125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef503fd02adc475fa79dda8b379b6514", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c964bdc6-fccc-40d9-bfe2-763b6f05a863", "external-id": "cl2-zone-376", "segmentation_id": 376, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69314b3d-cb", "ovs_interfaceid": "69314b3d-cb79-40f5-8562-afed90be617f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.277744] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293761, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505493} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.277744] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 9683894b-a300-4400-a1b9-db62478f42c5/9683894b-a300-4400-a1b9-db62478f42c5.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 848.277744] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 848.278038] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27b803ac-462c-4afb-b815-5179d989dae7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.284578] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 848.284578] env[62814]: value = "task-4293762" [ 848.284578] env[62814]: _type = "Task" [ 848.284578] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.293104] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293762, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.489832] env[62814]: DEBUG nova.network.neutron [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Successfully created port: aa760165-654c-49b6-97bb-5fc792c2071d {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 848.518200] env[62814]: DEBUG nova.compute.manager [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 848.562403] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293760, 'name': CreateSnapshot_Task, 'duration_secs': 0.63908} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.565608] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 848.566626] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8210ac2-f530-4043-804a-16f3ae05dfba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.614367] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe13801-3aaf-4f87-88b1-c94203cd8b6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.618106] env[62814]: DEBUG nova.network.neutron [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 848.620971] env[62814]: DEBUG oslo_concurrency.lockutils [req-7b15135d-f67b-4c59-b6ea-c7b5ef198873 req-83964850-d43d-4b8f-a670-9577226dc27c service nova] Releasing lock "refresh_cache-425e8edd-c002-45a1-bb6f-ee3ac8812509" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 848.627427] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0906a6d-653d-43b0-a2c2-feeb16ee8ab9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.661735] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e580b6a-cc95-45d4-a222-aa956c10047a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.670046] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e6caae-dc85-424a-9eea-3be5082c8af5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.685759] env[62814]: DEBUG nova.compute.provider_tree [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.795457] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293762, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057491} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.796325] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.796507] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e5b4f66-34af-4e92-bb0f-d27c14c42b97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.818354] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 9683894b-a300-4400-a1b9-db62478f42c5/9683894b-a300-4400-a1b9-db62478f42c5.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.818639] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5998735f-cf40-4c62-b90b-543f94e5cfbc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.834488] env[62814]: DEBUG nova.network.neutron [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Updating instance_info_cache with network_info: [{"id": "b6e71b20-cd42-432e-bc6d-1c16b443ed92", "address": "fa:16:3e:46:a8:b7", "network": {"id": "27b010b3-f6fb-452f-a43b-653ffbc67eba", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1286101218-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f39e30833c0d4ed8bd3e3870ab9084d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89ed4797-90ad-44cd-bbcb-e90b2a8400f3", "external-id": "nsx-vlan-transportzone-699", "segmentation_id": 699, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e71b20-cd", "ovs_interfaceid": "b6e71b20-cd42-432e-bc6d-1c16b443ed92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.838391] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 848.838391] env[62814]: value = "task-4293763" [ 848.838391] env[62814]: _type = "Task" [ 848.838391] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.847945] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293763, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.007590] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.007785] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.087033] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 849.087356] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d4f26076-6fbf-4f07-bbc2-d82775e56d1f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.097433] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 849.097433] env[62814]: value = "task-4293764" [ 849.097433] env[62814]: _type = "Task" [ 849.097433] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.108221] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293764, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.191505] env[62814]: DEBUG nova.scheduler.client.report [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 849.337366] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Releasing lock "refresh_cache-448b7bfb-1a90-4c5b-9086-9320cedaf015" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 849.337722] env[62814]: DEBUG nova.compute.manager [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Instance network_info: |[{"id": "b6e71b20-cd42-432e-bc6d-1c16b443ed92", "address": "fa:16:3e:46:a8:b7", "network": {"id": "27b010b3-f6fb-452f-a43b-653ffbc67eba", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1286101218-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f39e30833c0d4ed8bd3e3870ab9084d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89ed4797-90ad-44cd-bbcb-e90b2a8400f3", "external-id": "nsx-vlan-transportzone-699", "segmentation_id": 699, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e71b20-cd", "ovs_interfaceid": "b6e71b20-cd42-432e-bc6d-1c16b443ed92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 849.342963] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:a8:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '89ed4797-90ad-44cd-bbcb-e90b2a8400f3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6e71b20-cd42-432e-bc6d-1c16b443ed92', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.350496] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Creating folder: Project (f39e30833c0d4ed8bd3e3870ab9084d1). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.351585] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-210b28e6-8109-4112-a261-63f4579cfcd3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.365995] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293763, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.368168] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Created folder: Project (f39e30833c0d4ed8bd3e3870ab9084d1) in parent group-v845547. [ 849.368349] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Creating folder: Instances. Parent ref: group-v845700. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.368620] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ab61a92-e208-492c-8e2e-61b3fdd6670a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.377267] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Created folder: Instances in parent group-v845700. [ 849.377501] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 849.377689] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.377946] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-78841200-26ee-491d-8a00-dd2f37cfc49e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.397363] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.397363] env[62814]: value = "task-4293767" [ 849.397363] env[62814]: _type = "Task" [ 849.397363] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.405137] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293767, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.510542] env[62814]: DEBUG nova.compute.manager [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 849.528579] env[62814]: DEBUG nova.compute.manager [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 849.540555] env[62814]: DEBUG nova.compute.manager [req-ea05440d-c09c-4dd5-936b-503a53675a72 req-247bb4a3-4020-4979-9899-7d7208b201d8 service nova] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Received event network-changed-b6e71b20-cd42-432e-bc6d-1c16b443ed92 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 849.540623] env[62814]: DEBUG nova.compute.manager [req-ea05440d-c09c-4dd5-936b-503a53675a72 req-247bb4a3-4020-4979-9899-7d7208b201d8 service nova] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Refreshing instance network info cache due to event network-changed-b6e71b20-cd42-432e-bc6d-1c16b443ed92. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 849.540923] env[62814]: DEBUG oslo_concurrency.lockutils [req-ea05440d-c09c-4dd5-936b-503a53675a72 req-247bb4a3-4020-4979-9899-7d7208b201d8 service nova] Acquiring lock "refresh_cache-448b7bfb-1a90-4c5b-9086-9320cedaf015" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.541041] env[62814]: DEBUG oslo_concurrency.lockutils [req-ea05440d-c09c-4dd5-936b-503a53675a72 req-247bb4a3-4020-4979-9899-7d7208b201d8 service nova] Acquired lock "refresh_cache-448b7bfb-1a90-4c5b-9086-9320cedaf015" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 849.541378] env[62814]: DEBUG nova.network.neutron [req-ea05440d-c09c-4dd5-936b-503a53675a72 req-247bb4a3-4020-4979-9899-7d7208b201d8 service nova] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Refreshing network info cache for port b6e71b20-cd42-432e-bc6d-1c16b443ed92 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.553257] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 849.553510] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.553669] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 849.555866] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.555866] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 849.555866] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 849.555866] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 849.555866] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 849.555866] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 849.555866] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 849.555866] env[62814]: DEBUG nova.virt.hardware [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 849.556752] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2272b0-9d4a-400f-90f0-7b76fd93b86c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.566535] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52de9755-fb65-4167-8c63-88fe7adc60fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.606718] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293764, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.698062] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.192s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.699584] env[62814]: DEBUG oslo_concurrency.lockutils [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 36.329s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.699856] env[62814]: DEBUG nova.objects.instance [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lazy-loading 'resources' on Instance uuid 9677b03f-7138-47b7-b1e8-f3714d11e550 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.727451] env[62814]: INFO nova.scheduler.client.report [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Deleted allocations for instance 127fee64-fd56-4a23-bdd2-18c817898fd5 [ 849.852628] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquiring lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.852628] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.853332] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquiring lock "3d3f32fc-276a-49be-b471-01a5d6fc5069-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.853332] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.853332] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.859682] env[62814]: INFO nova.compute.manager [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Terminating instance [ 849.869554] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293763, 'name': ReconfigVM_Task, 'duration_secs': 0.780473} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.870423] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 9683894b-a300-4400-a1b9-db62478f42c5/9683894b-a300-4400-a1b9-db62478f42c5.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.870580] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cf04f2b3-820b-46f1-bf1c-9ec00124715d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.880083] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 849.880083] env[62814]: value = "task-4293768" [ 849.880083] env[62814]: _type = "Task" [ 849.880083] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.892228] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293768, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.909137] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293767, 'name': CreateVM_Task, 'duration_secs': 0.329233} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.909341] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 849.910036] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.910234] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 849.910576] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 849.911160] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d535d57d-ed42-4378-ab68-27ef11cc261d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.916464] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for the task: (returnval){ [ 849.916464] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d236f1-73ba-56fc-ad62-bdc8af60c6a4" [ 849.916464] env[62814]: _type = "Task" [ 849.916464] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.925785] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d236f1-73ba-56fc-ad62-bdc8af60c6a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.019893] env[62814]: DEBUG nova.compute.manager [req-f58337f5-cd63-4ac6-ab27-50e87dc20aad req-e4086354-d528-4b6d-8495-d7c340bdf2cb service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Received event network-vif-plugged-aa760165-654c-49b6-97bb-5fc792c2071d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 850.020268] env[62814]: DEBUG oslo_concurrency.lockutils [req-f58337f5-cd63-4ac6-ab27-50e87dc20aad req-e4086354-d528-4b6d-8495-d7c340bdf2cb service nova] Acquiring lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.020674] env[62814]: DEBUG oslo_concurrency.lockutils [req-f58337f5-cd63-4ac6-ab27-50e87dc20aad req-e4086354-d528-4b6d-8495-d7c340bdf2cb service nova] Lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.020898] env[62814]: DEBUG oslo_concurrency.lockutils [req-f58337f5-cd63-4ac6-ab27-50e87dc20aad req-e4086354-d528-4b6d-8495-d7c340bdf2cb service nova] Lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.021037] env[62814]: DEBUG nova.compute.manager [req-f58337f5-cd63-4ac6-ab27-50e87dc20aad req-e4086354-d528-4b6d-8495-d7c340bdf2cb service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] No waiting events found dispatching network-vif-plugged-aa760165-654c-49b6-97bb-5fc792c2071d {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 850.021408] env[62814]: WARNING nova.compute.manager [req-f58337f5-cd63-4ac6-ab27-50e87dc20aad req-e4086354-d528-4b6d-8495-d7c340bdf2cb service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Received unexpected event network-vif-plugged-aa760165-654c-49b6-97bb-5fc792c2071d for instance with vm_state building and task_state spawning. [ 850.043481] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.107932] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293764, 'name': CloneVM_Task} progress is 95%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.131526] env[62814]: DEBUG nova.network.neutron [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Successfully updated port: aa760165-654c-49b6-97bb-5fc792c2071d {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 850.236304] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e90752a2-fc41-4917-bf5d-ef0a4e585991 tempest-ServersListShow2100Test-1897042634 tempest-ServersListShow2100Test-1897042634-project-member] Lock "127fee64-fd56-4a23-bdd2-18c817898fd5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.649s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.365404] env[62814]: DEBUG nova.compute.manager [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 850.365880] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 850.369169] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f5ba6b-34b0-49cf-b650-d7349a74e111 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.383022] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 850.386153] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8483c5ae-a30f-4b2d-8240-0478d426813a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.395909] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293768, 'name': Rename_Task, 'duration_secs': 0.191752} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.397166] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 850.397576] env[62814]: DEBUG oslo_vmware.api [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for the task: (returnval){ [ 850.397576] env[62814]: value = "task-4293769" [ 850.397576] env[62814]: _type = "Task" [ 850.397576] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.397866] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-852b53c1-258c-4bc5-8662-35f0dbba94cf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.407434] env[62814]: DEBUG nova.network.neutron [req-ea05440d-c09c-4dd5-936b-503a53675a72 req-247bb4a3-4020-4979-9899-7d7208b201d8 service nova] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Updated VIF entry in instance network info cache for port b6e71b20-cd42-432e-bc6d-1c16b443ed92. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 850.407950] env[62814]: DEBUG nova.network.neutron [req-ea05440d-c09c-4dd5-936b-503a53675a72 req-247bb4a3-4020-4979-9899-7d7208b201d8 service nova] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Updating instance_info_cache with network_info: [{"id": "b6e71b20-cd42-432e-bc6d-1c16b443ed92", "address": "fa:16:3e:46:a8:b7", "network": {"id": "27b010b3-f6fb-452f-a43b-653ffbc67eba", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1286101218-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f39e30833c0d4ed8bd3e3870ab9084d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "89ed4797-90ad-44cd-bbcb-e90b2a8400f3", "external-id": "nsx-vlan-transportzone-699", "segmentation_id": 699, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6e71b20-cd", "ovs_interfaceid": "b6e71b20-cd42-432e-bc6d-1c16b443ed92", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.412556] env[62814]: DEBUG oslo_vmware.api [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293769, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.414161] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 850.414161] env[62814]: value = "task-4293770" [ 850.414161] env[62814]: _type = "Task" [ 850.414161] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.430886] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293770, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.434303] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d236f1-73ba-56fc-ad62-bdc8af60c6a4, 'name': SearchDatastore_Task, 'duration_secs': 0.011301} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.437391] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 850.437654] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.437905] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.438127] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 850.438327] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.438776] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-58f9eed0-8d79-4124-82ba-a595b7299e68 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.447360] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.447597] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.448640] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-203f948a-3ef9-4dff-a7e0-d2b0422f8487 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.458506] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for the task: (returnval){ [ 850.458506] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52400c05-4f64-b9ab-f400-110999d4ea11" [ 850.458506] env[62814]: _type = "Task" [ 850.458506] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.467697] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52400c05-4f64-b9ab-f400-110999d4ea11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.613835] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293764, 'name': CloneVM_Task, 'duration_secs': 1.137439} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.614318] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Created linked-clone VM from snapshot [ 850.615651] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834756e5-4fec-4f5a-8935-a3652d13c101 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.630430] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Uploading image 577d3b83-11e9-430b-a3da-3fc2f382fd79 {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 850.634316] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "refresh_cache-ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.634463] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "refresh_cache-ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 850.634612] env[62814]: DEBUG nova.network.neutron [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 850.665572] env[62814]: DEBUG oslo_vmware.rw_handles [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 850.665572] env[62814]: value = "vm-845699" [ 850.665572] env[62814]: _type = "VirtualMachine" [ 850.665572] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 850.665572] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a494fd25-e7cc-4be0-8928-f27ea91da52f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.673172] env[62814]: DEBUG oslo_vmware.rw_handles [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lease: (returnval){ [ 850.673172] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528bd6fd-d030-e538-a25c-53c1fd696c9f" [ 850.673172] env[62814]: _type = "HttpNfcLease" [ 850.673172] env[62814]: } obtained for exporting VM: (result){ [ 850.673172] env[62814]: value = "vm-845699" [ 850.673172] env[62814]: _type = "VirtualMachine" [ 850.673172] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 850.673451] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the lease: (returnval){ [ 850.673451] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528bd6fd-d030-e538-a25c-53c1fd696c9f" [ 850.673451] env[62814]: _type = "HttpNfcLease" [ 850.673451] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 850.685471] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 850.685471] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528bd6fd-d030-e538-a25c-53c1fd696c9f" [ 850.685471] env[62814]: _type = "HttpNfcLease" [ 850.685471] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 850.767904] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc0beb2-3cf8-41f9-ba08-f77cd526ca01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.778622] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a877354c-25db-4433-a791-e28cd352273d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.812588] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6516167c-ba85-463c-880e-60ed666a928b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.820896] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e625f8-2eb2-4d98-a427-1aa4d1e8fedc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.837609] env[62814]: DEBUG nova.compute.provider_tree [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.914553] env[62814]: DEBUG oslo_concurrency.lockutils [req-ea05440d-c09c-4dd5-936b-503a53675a72 req-247bb4a3-4020-4979-9899-7d7208b201d8 service nova] Releasing lock "refresh_cache-448b7bfb-1a90-4c5b-9086-9320cedaf015" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 850.914910] env[62814]: DEBUG oslo_vmware.api [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293769, 'name': PowerOffVM_Task, 'duration_secs': 0.213824} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.915271] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 850.915517] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 850.919414] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-22a013d1-4cca-496a-92b6-f79caa8bfd0a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.927254] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293770, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.970349] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52400c05-4f64-b9ab-f400-110999d4ea11, 'name': SearchDatastore_Task, 'duration_secs': 0.010183} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.971620] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8423c593-76e1-4754-980d-fb9d0d98a5b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.976017] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 850.976685] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 850.976685] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Deleting the datastore file [datastore2] 3d3f32fc-276a-49be-b471-01a5d6fc5069 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 850.976865] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b1004cf3-b749-449c-83eb-cb4e60a25ce3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.980773] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for the task: (returnval){ [ 850.980773] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529eebef-7ff6-69dd-e455-caf4eef391aa" [ 850.980773] env[62814]: _type = "Task" [ 850.980773] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.986576] env[62814]: DEBUG oslo_vmware.api [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for the task: (returnval){ [ 850.986576] env[62814]: value = "task-4293773" [ 850.986576] env[62814]: _type = "Task" [ 850.986576] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.990407] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529eebef-7ff6-69dd-e455-caf4eef391aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.002672] env[62814]: DEBUG oslo_vmware.api [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293773, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.181355] env[62814]: DEBUG nova.network.neutron [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.187886] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 851.187886] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528bd6fd-d030-e538-a25c-53c1fd696c9f" [ 851.187886] env[62814]: _type = "HttpNfcLease" [ 851.187886] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 851.188373] env[62814]: DEBUG oslo_vmware.rw_handles [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 851.188373] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528bd6fd-d030-e538-a25c-53c1fd696c9f" [ 851.188373] env[62814]: _type = "HttpNfcLease" [ 851.188373] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 851.189212] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca2bd35-d818-4f01-a114-13b511b67f98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.199569] env[62814]: DEBUG oslo_vmware.rw_handles [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527307af-9c2e-611e-45b8-670e7f9c168f/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 851.200081] env[62814]: DEBUG oslo_vmware.rw_handles [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527307af-9c2e-611e-45b8-670e7f9c168f/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 851.329814] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-14a3c1af-3470-46ac-93b3-44618e9276a3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.341680] env[62814]: DEBUG nova.scheduler.client.report [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 851.428580] env[62814]: DEBUG oslo_vmware.api [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4293770, 'name': PowerOnVM_Task, 'duration_secs': 0.704218} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.432694] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.432996] env[62814]: INFO nova.compute.manager [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Took 8.17 seconds to spawn the instance on the hypervisor. [ 851.433266] env[62814]: DEBUG nova.compute.manager [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 851.434420] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e26101-90cf-4111-b669-469731f276fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.445590] env[62814]: DEBUG nova.network.neutron [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Updating instance_info_cache with network_info: [{"id": "aa760165-654c-49b6-97bb-5fc792c2071d", "address": "fa:16:3e:34:a7:13", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa760165-65", "ovs_interfaceid": "aa760165-654c-49b6-97bb-5fc792c2071d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.492617] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529eebef-7ff6-69dd-e455-caf4eef391aa, 'name': SearchDatastore_Task, 'duration_secs': 0.010565} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.496503] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 851.496949] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 448b7bfb-1a90-4c5b-9086-9320cedaf015/448b7bfb-1a90-4c5b-9086-9320cedaf015.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.497570] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7dd100ed-b858-41b6-b071-c5fedbc7f8ee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.505351] env[62814]: DEBUG oslo_vmware.api [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Task: {'id': task-4293773, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150735} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.507331] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 851.507331] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 851.507331] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 851.507608] env[62814]: INFO nova.compute.manager [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Took 1.14 seconds to destroy the instance on the hypervisor. [ 851.507964] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 851.508331] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for the task: (returnval){ [ 851.508331] env[62814]: value = "task-4293774" [ 851.508331] env[62814]: _type = "Task" [ 851.508331] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.508608] env[62814]: DEBUG nova.compute.manager [-] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 851.509774] env[62814]: DEBUG nova.network.neutron [-] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 851.521907] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293774, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.847816] env[62814]: DEBUG oslo_concurrency.lockutils [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.148s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 851.851982] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.013s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 851.853967] env[62814]: INFO nova.compute.claims [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.901800] env[62814]: INFO nova.scheduler.client.report [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Deleted allocations for instance 9677b03f-7138-47b7-b1e8-f3714d11e550 [ 851.953615] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "refresh_cache-ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 851.954096] env[62814]: DEBUG nova.compute.manager [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Instance network_info: |[{"id": "aa760165-654c-49b6-97bb-5fc792c2071d", "address": "fa:16:3e:34:a7:13", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa760165-65", "ovs_interfaceid": "aa760165-654c-49b6-97bb-5fc792c2071d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 851.957122] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:34:a7:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa760165-654c-49b6-97bb-5fc792c2071d', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.969860] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 851.969860] env[62814]: INFO nova.compute.manager [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Took 49.74 seconds to build instance. [ 851.969860] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 851.969860] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2c9f1c2-e3d3-4f3c-8841-a54d407b73cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.991046] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.991046] env[62814]: value = "task-4293775" [ 851.991046] env[62814]: _type = "Task" [ 851.991046] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.998464] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293775, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.026957] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293774, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.086469] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 852.086929] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 852.197642] env[62814]: DEBUG nova.compute.manager [req-4a28ab52-1525-40bb-93d3-3a28850c260b req-442842a0-6630-4582-8f1d-a0a80355d0a3 service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Received event network-changed-aa760165-654c-49b6-97bb-5fc792c2071d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 852.197959] env[62814]: DEBUG nova.compute.manager [req-4a28ab52-1525-40bb-93d3-3a28850c260b req-442842a0-6630-4582-8f1d-a0a80355d0a3 service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Refreshing instance network info cache due to event network-changed-aa760165-654c-49b6-97bb-5fc792c2071d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 852.200068] env[62814]: DEBUG oslo_concurrency.lockutils [req-4a28ab52-1525-40bb-93d3-3a28850c260b req-442842a0-6630-4582-8f1d-a0a80355d0a3 service nova] Acquiring lock "refresh_cache-ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.200302] env[62814]: DEBUG oslo_concurrency.lockutils [req-4a28ab52-1525-40bb-93d3-3a28850c260b req-442842a0-6630-4582-8f1d-a0a80355d0a3 service nova] Acquired lock "refresh_cache-ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 852.200474] env[62814]: DEBUG nova.network.neutron [req-4a28ab52-1525-40bb-93d3-3a28850c260b req-442842a0-6630-4582-8f1d-a0a80355d0a3 service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Refreshing network info cache for port aa760165-654c-49b6-97bb-5fc792c2071d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 852.412376] env[62814]: DEBUG oslo_concurrency.lockutils [None req-951707bc-580d-4c13-9b4d-ece3ecacdd80 tempest-ImagesOneServerTestJSON-1365087924 tempest-ImagesOneServerTestJSON-1365087924-project-member] Lock "9677b03f-7138-47b7-b1e8-f3714d11e550" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 42.590s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.485294] env[62814]: DEBUG oslo_concurrency.lockutils [None req-690f084f-8403-4373-aa61-38a2143547be tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.599s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.503100] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293775, 'name': CreateVM_Task, 'duration_secs': 0.438206} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.503512] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 852.505126] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.505290] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 852.505663] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 852.506598] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ce115a9-33f7-43fb-905e-2e11378fe5c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.512962] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 852.512962] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529c166b-4839-1454-bcbe-ba8eedd51e41" [ 852.512962] env[62814]: _type = "Task" [ 852.512962] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.531616] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529c166b-4839-1454-bcbe-ba8eedd51e41, 'name': SearchDatastore_Task, 'duration_secs': 0.009961} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.531891] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293774, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.54685} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.532153] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 852.532391] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.532716] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.533064] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 852.534127] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.534127] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 448b7bfb-1a90-4c5b-9086-9320cedaf015/448b7bfb-1a90-4c5b-9086-9320cedaf015.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.534127] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.534127] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0042e620-d77d-4570-8d72-980c3b774102 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.537308] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-950b482a-c2b6-49ea-852f-f03af4b1effd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.545446] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for the task: (returnval){ [ 852.545446] env[62814]: value = "task-4293776" [ 852.545446] env[62814]: _type = "Task" [ 852.545446] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.546900] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.547477] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 852.556954] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c092881c-3016-4424-b8e8-58ed7233c394 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.561329] env[62814]: DEBUG nova.compute.manager [req-33058276-b9d7-429c-9b79-d2df23932236 req-f727e9a4-779c-47a7-b5e4-0733f3a5fee7 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Received event network-vif-deleted-25fab244-2cae-416b-86e1-f25d1ea5aae7 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 852.561533] env[62814]: INFO nova.compute.manager [req-33058276-b9d7-429c-9b79-d2df23932236 req-f727e9a4-779c-47a7-b5e4-0733f3a5fee7 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Neutron deleted interface 25fab244-2cae-416b-86e1-f25d1ea5aae7; detaching it from the instance and deleting it from the info cache [ 852.561720] env[62814]: DEBUG nova.network.neutron [req-33058276-b9d7-429c-9b79-d2df23932236 req-f727e9a4-779c-47a7-b5e4-0733f3a5fee7 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.573763] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 852.573763] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b02ed-63d7-1788-4e94-64d6c94145d4" [ 852.573763] env[62814]: _type = "Task" [ 852.573763] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.574077] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293776, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.585283] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b02ed-63d7-1788-4e94-64d6c94145d4, 'name': SearchDatastore_Task, 'duration_secs': 0.010549} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.586569] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34cc312a-c5a3-40f4-8f81-bcca89d8d6e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.592602] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 852.592602] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520b399a-9fdd-1fad-4539-3189e365b5cc" [ 852.592602] env[62814]: _type = "Task" [ 852.592602] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.602656] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520b399a-9fdd-1fad-4539-3189e365b5cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.883601] env[62814]: DEBUG nova.network.neutron [-] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.989777] env[62814]: DEBUG nova.compute.manager [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 853.059157] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293776, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074458} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.059483] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.060362] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec63b17f-3b1f-4544-9c4f-03c34170d53c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.066570] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-01137b36-90ad-42f6-8cf4-b83e4acc53aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.087045] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 448b7bfb-1a90-4c5b-9086-9320cedaf015/448b7bfb-1a90-4c5b-9086-9320cedaf015.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.090773] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7b61b35-3f04-4acb-95c4-ac8381882d30 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.113392] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7879afe3-0d91-4ff1-a4c6-244e85609c02 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.128674] env[62814]: DEBUG nova.network.neutron [req-4a28ab52-1525-40bb-93d3-3a28850c260b req-442842a0-6630-4582-8f1d-a0a80355d0a3 service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Updated VIF entry in instance network info cache for port aa760165-654c-49b6-97bb-5fc792c2071d. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 853.129044] env[62814]: DEBUG nova.network.neutron [req-4a28ab52-1525-40bb-93d3-3a28850c260b req-442842a0-6630-4582-8f1d-a0a80355d0a3 service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Updating instance_info_cache with network_info: [{"id": "aa760165-654c-49b6-97bb-5fc792c2071d", "address": "fa:16:3e:34:a7:13", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa760165-65", "ovs_interfaceid": "aa760165-654c-49b6-97bb-5fc792c2071d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.130785] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for the task: (returnval){ [ 853.130785] env[62814]: value = "task-4293777" [ 853.130785] env[62814]: _type = "Task" [ 853.130785] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.140595] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520b399a-9fdd-1fad-4539-3189e365b5cc, 'name': SearchDatastore_Task, 'duration_secs': 0.009891} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.141629] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 853.141840] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01/ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 853.157493] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9388ae6-1db9-47b9-b6d9-ceaf17ac8da3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.159952] env[62814]: DEBUG nova.compute.manager [req-33058276-b9d7-429c-9b79-d2df23932236 req-f727e9a4-779c-47a7-b5e4-0733f3a5fee7 service nova] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Detach interface failed, port_id=25fab244-2cae-416b-86e1-f25d1ea5aae7, reason: Instance 3d3f32fc-276a-49be-b471-01a5d6fc5069 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 853.170977] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293777, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.172400] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "d7332882-19b3-4ab3-8ea6-51d33c584844" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.172914] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "d7332882-19b3-4ab3-8ea6-51d33c584844" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 853.177229] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 853.177229] env[62814]: value = "task-4293778" [ 853.177229] env[62814]: _type = "Task" [ 853.177229] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.186431] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293778, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.385943] env[62814]: INFO nova.compute.manager [-] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Took 1.88 seconds to deallocate network for instance. [ 853.520337] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 853.559330] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c732fe1-f60e-4f68-9a1f-c3ec5231dcfa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.567644] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f45429-2704-4c84-a1fe-39e6ed22bde1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.600613] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8963edbd-9515-4f37-b25c-1dbd9baf7858 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.609497] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233ed366-0357-4118-9d5d-c6cfd21fdd8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.624270] env[62814]: DEBUG nova.compute.provider_tree [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.633062] env[62814]: DEBUG oslo_concurrency.lockutils [req-4a28ab52-1525-40bb-93d3-3a28850c260b req-442842a0-6630-4582-8f1d-a0a80355d0a3 service nova] Releasing lock "refresh_cache-ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 853.645891] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293777, 'name': ReconfigVM_Task, 'duration_secs': 0.377369} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.645891] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 448b7bfb-1a90-4c5b-9086-9320cedaf015/448b7bfb-1a90-4c5b-9086-9320cedaf015.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.645891] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f59faba7-93a5-4a26-8ca3-c949ef1c4947 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.652275] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for the task: (returnval){ [ 853.652275] env[62814]: value = "task-4293779" [ 853.652275] env[62814]: _type = "Task" [ 853.652275] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.661219] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293779, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.686689] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293778, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.896295] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 854.129880] env[62814]: DEBUG nova.scheduler.client.report [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 854.164776] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293779, 'name': Rename_Task, 'duration_secs': 0.253422} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.165210] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 854.165585] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-677d53fd-9d26-4460-bd05-de2ce08a60b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.174031] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for the task: (returnval){ [ 854.174031] env[62814]: value = "task-4293780" [ 854.174031] env[62814]: _type = "Task" [ 854.174031] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.188459] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.192898] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293778, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.594927} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.193269] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01/ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 854.193563] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.193877] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a647d4a-ba9e-43da-a50b-382837f9f5ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.199783] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 854.199783] env[62814]: value = "task-4293781" [ 854.199783] env[62814]: _type = "Task" [ 854.199783] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.209913] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293781, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.231486] env[62814]: DEBUG nova.compute.manager [req-040b96fd-c87f-45bc-a70e-d0064e87a2cf req-8b54b112-459a-4fd1-9143-1356a44ff719 service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Received event network-changed-9fed46ec-04e6-4482-a460-370e4ff04a32 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 854.231677] env[62814]: DEBUG nova.compute.manager [req-040b96fd-c87f-45bc-a70e-d0064e87a2cf req-8b54b112-459a-4fd1-9143-1356a44ff719 service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Refreshing instance network info cache due to event network-changed-9fed46ec-04e6-4482-a460-370e4ff04a32. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 854.231899] env[62814]: DEBUG oslo_concurrency.lockutils [req-040b96fd-c87f-45bc-a70e-d0064e87a2cf req-8b54b112-459a-4fd1-9143-1356a44ff719 service nova] Acquiring lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.232324] env[62814]: DEBUG oslo_concurrency.lockutils [req-040b96fd-c87f-45bc-a70e-d0064e87a2cf req-8b54b112-459a-4fd1-9143-1356a44ff719 service nova] Acquired lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 854.233378] env[62814]: DEBUG nova.network.neutron [req-040b96fd-c87f-45bc-a70e-d0064e87a2cf req-8b54b112-459a-4fd1-9143-1356a44ff719 service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Refreshing network info cache for port 9fed46ec-04e6-4482-a460-370e4ff04a32 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.633256] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.781s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 854.633951] env[62814]: DEBUG nova.compute.manager [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 854.638305] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 40.607s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 854.639983] env[62814]: INFO nova.compute.claims [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.690704] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293780, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.711522] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293781, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095388} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.711648] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.712556] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5e19a9-4853-4599-9ae3-f5eb51c8d019 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.750700] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01/ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.751502] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-420085ab-c284-4159-bc81-e891fbd04f24 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.781094] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 854.781094] env[62814]: value = "task-4293782" [ 854.781094] env[62814]: _type = "Task" [ 854.781094] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.790580] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293782, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.153744] env[62814]: DEBUG nova.compute.utils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 855.154937] env[62814]: DEBUG nova.compute.manager [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 855.155112] env[62814]: DEBUG nova.network.neutron [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 855.184680] env[62814]: DEBUG oslo_vmware.api [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293780, 'name': PowerOnVM_Task, 'duration_secs': 0.60508} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.184960] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 855.185686] env[62814]: INFO nova.compute.manager [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Took 8.28 seconds to spawn the instance on the hypervisor. [ 855.186033] env[62814]: DEBUG nova.compute.manager [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 855.187740] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94d03f6-d99e-45b3-8b60-bd090f7fea6e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.194807] env[62814]: DEBUG nova.network.neutron [req-040b96fd-c87f-45bc-a70e-d0064e87a2cf req-8b54b112-459a-4fd1-9143-1356a44ff719 service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Updated VIF entry in instance network info cache for port 9fed46ec-04e6-4482-a460-370e4ff04a32. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.195109] env[62814]: DEBUG nova.network.neutron [req-040b96fd-c87f-45bc-a70e-d0064e87a2cf req-8b54b112-459a-4fd1-9143-1356a44ff719 service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Updating instance_info_cache with network_info: [{"id": "9fed46ec-04e6-4482-a460-370e4ff04a32", "address": "fa:16:3e:19:a2:ac", "network": {"id": "5363651c-6ade-4a4c-b762-9fc7cbbcf5ea", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1833251571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b3ba1c9c7624abfb3e0bd880dc40166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fed46ec-04", "ovs_interfaceid": "9fed46ec-04e6-4482-a460-370e4ff04a32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.205099] env[62814]: DEBUG nova.policy [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2b663046f1c5483a9b12968b16ccd98b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd14adacd89bc460d8d1a7c7d2ecc5d7f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 855.291557] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293782, 'name': ReconfigVM_Task, 'duration_secs': 0.359053} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.291863] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Reconfigured VM instance instance-00000037 to attach disk [datastore2] ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01/ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.292574] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d6faf86-d1c2-445e-b649-ba6856ac2092 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.299970] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 855.299970] env[62814]: value = "task-4293783" [ 855.299970] env[62814]: _type = "Task" [ 855.299970] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.311479] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293783, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.540392] env[62814]: DEBUG nova.network.neutron [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Successfully created port: 526f5c77-efba-4403-b12e-cc66f3cbd46d {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.662459] env[62814]: DEBUG nova.compute.manager [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 855.708070] env[62814]: DEBUG oslo_concurrency.lockutils [req-040b96fd-c87f-45bc-a70e-d0064e87a2cf req-8b54b112-459a-4fd1-9143-1356a44ff719 service nova] Releasing lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 855.716209] env[62814]: INFO nova.compute.manager [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Took 50.25 seconds to build instance. [ 855.812989] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293783, 'name': Rename_Task, 'duration_secs': 0.167608} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.812989] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.813250] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b763e8c0-d7dc-441e-9502-d8efd8438f99 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.820306] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 855.820306] env[62814]: value = "task-4293784" [ 855.820306] env[62814]: _type = "Task" [ 855.820306] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.832361] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293784, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.990794] env[62814]: DEBUG nova.network.neutron [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Successfully created port: 5d635de7-0d82-4d28-a72c-35736e7d8f3b {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 856.122864] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquiring lock "448b7bfb-1a90-4c5b-9086-9320cedaf015" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.218362] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62e602e7-7d89-4ff2-878f-8747652b95cf tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "448b7bfb-1a90-4c5b-9086-9320cedaf015" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.578s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.220165] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "448b7bfb-1a90-4c5b-9086-9320cedaf015" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.097s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.220532] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquiring lock "448b7bfb-1a90-4c5b-9086-9320cedaf015-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.220871] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "448b7bfb-1a90-4c5b-9086-9320cedaf015-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.221292] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "448b7bfb-1a90-4c5b-9086-9320cedaf015-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.224789] env[62814]: INFO nova.compute.manager [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Terminating instance [ 856.293956] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82d5488-8c50-4aa8-ab36-8e97d8eded8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.302604] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d88a2d-61f0-4287-a491-51e3d9a17816 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.339910] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96a2c6e-5c60-4bcb-ab6b-a44b0cb98c7e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.350032] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293784, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.351559] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a210f7-63fe-445e-88c8-f17f34f3862e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.366229] env[62814]: DEBUG nova.compute.provider_tree [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.671865] env[62814]: DEBUG nova.compute.manager [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 856.696536] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 856.696842] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.696999] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 856.697233] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.697439] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 856.697602] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 856.697815] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 856.697975] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 856.698182] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 856.698357] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 856.698532] env[62814]: DEBUG nova.virt.hardware [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 856.699382] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c229f661-ba36-4975-a57c-0559ab5c06b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.707365] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c456239-a8ce-4bff-8e83-480be29039c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.724354] env[62814]: DEBUG nova.compute.manager [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 856.728798] env[62814]: DEBUG nova.compute.manager [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 856.729026] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 856.729829] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd46b47-e63b-4563-918b-5ccf5519a61a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.739506] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 856.739846] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8a620a5-a45d-4e1d-9842-a719b690ded1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.749467] env[62814]: DEBUG oslo_vmware.api [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for the task: (returnval){ [ 856.749467] env[62814]: value = "task-4293785" [ 856.749467] env[62814]: _type = "Task" [ 856.749467] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.758427] env[62814]: DEBUG oslo_vmware.api [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293785, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.846070] env[62814]: DEBUG oslo_vmware.api [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293784, 'name': PowerOnVM_Task, 'duration_secs': 0.55807} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.846372] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.846615] env[62814]: INFO nova.compute.manager [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Took 7.32 seconds to spawn the instance on the hypervisor. [ 856.846799] env[62814]: DEBUG nova.compute.manager [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 856.847602] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4500c989-f735-46c9-9ea3-598010a03d27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.870735] env[62814]: DEBUG nova.scheduler.client.report [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 857.248528] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 857.257726] env[62814]: DEBUG oslo_concurrency.lockutils [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] Acquiring lock "refresh_cache-ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.257726] env[62814]: DEBUG oslo_concurrency.lockutils [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] Acquired lock "refresh_cache-ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 857.257926] env[62814]: DEBUG nova.network.neutron [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.264827] env[62814]: DEBUG oslo_vmware.api [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293785, 'name': PowerOffVM_Task, 'duration_secs': 0.216358} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.265095] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 857.265302] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 857.265564] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be7621f3-45b4-4c70-926f-f25b9b12ef4f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.332580] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 857.332830] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 857.333017] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Deleting the datastore file [datastore2] 448b7bfb-1a90-4c5b-9086-9320cedaf015 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 857.333278] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-953dc410-b088-4064-a718-9ea4cbc6c232 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.339854] env[62814]: DEBUG oslo_vmware.api [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for the task: (returnval){ [ 857.339854] env[62814]: value = "task-4293787" [ 857.339854] env[62814]: _type = "Task" [ 857.339854] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.348192] env[62814]: DEBUG oslo_vmware.api [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293787, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.363937] env[62814]: INFO nova.compute.manager [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Took 47.69 seconds to build instance. [ 857.375858] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.738s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.376384] env[62814]: DEBUG nova.compute.manager [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 857.379275] env[62814]: DEBUG oslo_concurrency.lockutils [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.042s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.379553] env[62814]: DEBUG nova.objects.instance [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lazy-loading 'resources' on Instance uuid 2d0b24c5-3593-4ef2-a637-d3590242ad79 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 857.789346] env[62814]: DEBUG nova.network.neutron [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Successfully updated port: 526f5c77-efba-4403-b12e-cc66f3cbd46d {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.855965] env[62814]: DEBUG oslo_vmware.api [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Task: {'id': task-4293787, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.418654} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.858223] env[62814]: DEBUG nova.compute.manager [req-1f318300-2e8a-4b92-8f11-9f0f22bc48f8 req-f0b441cc-83d8-4a75-b71e-bb39a9bde460 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Received event network-vif-plugged-526f5c77-efba-4403-b12e-cc66f3cbd46d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 857.858371] env[62814]: DEBUG oslo_concurrency.lockutils [req-1f318300-2e8a-4b92-8f11-9f0f22bc48f8 req-f0b441cc-83d8-4a75-b71e-bb39a9bde460 service nova] Acquiring lock "a473cc7e-0f86-4a18-9789-0aabfff430ab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 857.862209] env[62814]: DEBUG oslo_concurrency.lockutils [req-1f318300-2e8a-4b92-8f11-9f0f22bc48f8 req-f0b441cc-83d8-4a75-b71e-bb39a9bde460 service nova] Lock "a473cc7e-0f86-4a18-9789-0aabfff430ab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.862209] env[62814]: DEBUG oslo_concurrency.lockutils [req-1f318300-2e8a-4b92-8f11-9f0f22bc48f8 req-f0b441cc-83d8-4a75-b71e-bb39a9bde460 service nova] Lock "a473cc7e-0f86-4a18-9789-0aabfff430ab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.862209] env[62814]: DEBUG nova.compute.manager [req-1f318300-2e8a-4b92-8f11-9f0f22bc48f8 req-f0b441cc-83d8-4a75-b71e-bb39a9bde460 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] No waiting events found dispatching network-vif-plugged-526f5c77-efba-4403-b12e-cc66f3cbd46d {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 857.862209] env[62814]: WARNING nova.compute.manager [req-1f318300-2e8a-4b92-8f11-9f0f22bc48f8 req-f0b441cc-83d8-4a75-b71e-bb39a9bde460 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Received unexpected event network-vif-plugged-526f5c77-efba-4403-b12e-cc66f3cbd46d for instance with vm_state building and task_state spawning. [ 857.862209] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 857.862209] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 857.862209] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 857.862209] env[62814]: INFO nova.compute.manager [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Took 1.13 seconds to destroy the instance on the hypervisor. [ 857.862209] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 857.862209] env[62814]: DEBUG nova.compute.manager [-] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 857.862209] env[62814]: DEBUG nova.network.neutron [-] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 857.865823] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1d0124b3-e563-4a6f-80f2-7d44eb43719a tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.090s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.882900] env[62814]: DEBUG nova.compute.utils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 857.888718] env[62814]: DEBUG nova.compute.manager [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 857.891014] env[62814]: DEBUG nova.network.neutron [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 857.998304] env[62814]: DEBUG nova.policy [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76ee4d6ec27b4156afc391165f0b752e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7f19ca23f48c40d391e7348ebf008751', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 858.087064] env[62814]: DEBUG nova.network.neutron [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Updating instance_info_cache with network_info: [{"id": "aa760165-654c-49b6-97bb-5fc792c2071d", "address": "fa:16:3e:34:a7:13", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa760165-65", "ovs_interfaceid": "aa760165-654c-49b6-97bb-5fc792c2071d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.389529] env[62814]: DEBUG nova.compute.manager [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 858.474533] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446afd87-cd4d-4ec8-8874-34479d6de4ef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.483572] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f05401f-8ccc-437c-b68d-c994179944d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.522869] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9597d272-82ee-4bb8-98c5-5d8a803b52c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.530854] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ad11427-59cc-48c0-baf6-d54db1f502a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.545115] env[62814]: DEBUG nova.compute.provider_tree [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.590096] env[62814]: DEBUG oslo_concurrency.lockutils [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] Releasing lock "refresh_cache-ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 858.590395] env[62814]: DEBUG nova.compute.manager [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Inject network info {{(pid=62814) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7737}} [ 858.590689] env[62814]: DEBUG nova.compute.manager [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] network_info to inject: |[{"id": "aa760165-654c-49b6-97bb-5fc792c2071d", "address": "fa:16:3e:34:a7:13", "network": {"id": "e850eed5-2943-49de-9b80-f82ad717b9ef", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-124866519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f11d1ef620764fa4b1e2b718ac207a44", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ffcecdaa-a7b8-49fc-9371-dbdb7744688e", "external-id": "nsx-vlan-transportzone-994", "segmentation_id": 994, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa760165-65", "ovs_interfaceid": "aa760165-654c-49b6-97bb-5fc792c2071d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7738}} [ 858.596543] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Reconfiguring VM instance to set the machine id {{(pid=62814) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 858.596882] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3dd77efa-aba6-4147-8cf5-543046aa5c8e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.613309] env[62814]: DEBUG oslo_vmware.api [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] Waiting for the task: (returnval){ [ 858.613309] env[62814]: value = "task-4293788" [ 858.613309] env[62814]: _type = "Task" [ 858.613309] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.622691] env[62814]: DEBUG oslo_vmware.api [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] Task: {'id': task-4293788, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.779274] env[62814]: DEBUG nova.network.neutron [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Successfully created port: 696f3530-9858-47e0-be99-e223f03b58fb {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 859.048555] env[62814]: DEBUG nova.scheduler.client.report [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 859.125272] env[62814]: DEBUG oslo_vmware.api [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] Task: {'id': task-4293788, 'name': ReconfigVM_Task, 'duration_secs': 0.172072} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.125389] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-555b51f8-33df-4235-b57f-d2bf181cb30c tempest-ServersAdminTestJSON-1410409404 tempest-ServersAdminTestJSON-1410409404-project-admin] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Reconfigured VM instance to set the machine id {{(pid=62814) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 859.136450] env[62814]: DEBUG nova.network.neutron [-] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.401698] env[62814]: DEBUG nova.compute.manager [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 859.425395] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 859.425754] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.425956] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 859.426216] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.426449] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 859.426594] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 859.426847] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 859.427699] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 859.427699] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 859.427699] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 859.427860] env[62814]: DEBUG nova.virt.hardware [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 859.428969] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95e9490-331f-41e0-a0e4-c5059c95c0c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.438578] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a4b5b99-2c02-4528-af9c-64ae400a3704 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.555766] env[62814]: DEBUG oslo_concurrency.lockutils [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.175s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.557583] env[62814]: DEBUG oslo_concurrency.lockutils [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 42.974s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.558093] env[62814]: DEBUG nova.objects.instance [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lazy-loading 'resources' on Instance uuid c18acec7-cf95-4cdf-aa49-32419d364534 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 859.594060] env[62814]: INFO nova.scheduler.client.report [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleted allocations for instance 2d0b24c5-3593-4ef2-a637-d3590242ad79 [ 859.639621] env[62814]: INFO nova.compute.manager [-] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Took 1.78 seconds to deallocate network for instance. [ 859.877427] env[62814]: DEBUG nova.compute.manager [req-0d2a05ef-68c5-4908-bcbd-5ba32be3d03f req-8d399308-fcce-4e4c-8b59-d44aff18c339 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Received event network-vif-plugged-5d635de7-0d82-4d28-a72c-35736e7d8f3b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 859.877840] env[62814]: DEBUG oslo_concurrency.lockutils [req-0d2a05ef-68c5-4908-bcbd-5ba32be3d03f req-8d399308-fcce-4e4c-8b59-d44aff18c339 service nova] Acquiring lock "a473cc7e-0f86-4a18-9789-0aabfff430ab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 859.879598] env[62814]: DEBUG oslo_concurrency.lockutils [req-0d2a05ef-68c5-4908-bcbd-5ba32be3d03f req-8d399308-fcce-4e4c-8b59-d44aff18c339 service nova] Lock "a473cc7e-0f86-4a18-9789-0aabfff430ab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.879598] env[62814]: DEBUG oslo_concurrency.lockutils [req-0d2a05ef-68c5-4908-bcbd-5ba32be3d03f req-8d399308-fcce-4e4c-8b59-d44aff18c339 service nova] Lock "a473cc7e-0f86-4a18-9789-0aabfff430ab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 859.879598] env[62814]: DEBUG nova.compute.manager [req-0d2a05ef-68c5-4908-bcbd-5ba32be3d03f req-8d399308-fcce-4e4c-8b59-d44aff18c339 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] No waiting events found dispatching network-vif-plugged-5d635de7-0d82-4d28-a72c-35736e7d8f3b {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 859.879598] env[62814]: WARNING nova.compute.manager [req-0d2a05ef-68c5-4908-bcbd-5ba32be3d03f req-8d399308-fcce-4e4c-8b59-d44aff18c339 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Received unexpected event network-vif-plugged-5d635de7-0d82-4d28-a72c-35736e7d8f3b for instance with vm_state building and task_state spawning. [ 859.923483] env[62814]: DEBUG nova.compute.manager [req-4cc5beb4-9015-4713-bdff-f44ed887a50c req-698b76f3-ad1b-4795-aa14-f4f2885d8de1 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Received event network-changed-526f5c77-efba-4403-b12e-cc66f3cbd46d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 859.923483] env[62814]: DEBUG nova.compute.manager [req-4cc5beb4-9015-4713-bdff-f44ed887a50c req-698b76f3-ad1b-4795-aa14-f4f2885d8de1 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Refreshing instance network info cache due to event network-changed-526f5c77-efba-4403-b12e-cc66f3cbd46d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 859.923483] env[62814]: DEBUG oslo_concurrency.lockutils [req-4cc5beb4-9015-4713-bdff-f44ed887a50c req-698b76f3-ad1b-4795-aa14-f4f2885d8de1 service nova] Acquiring lock "refresh_cache-a473cc7e-0f86-4a18-9789-0aabfff430ab" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.923483] env[62814]: DEBUG oslo_concurrency.lockutils [req-4cc5beb4-9015-4713-bdff-f44ed887a50c req-698b76f3-ad1b-4795-aa14-f4f2885d8de1 service nova] Acquired lock "refresh_cache-a473cc7e-0f86-4a18-9789-0aabfff430ab" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 859.923483] env[62814]: DEBUG nova.network.neutron [req-4cc5beb4-9015-4713-bdff-f44ed887a50c req-698b76f3-ad1b-4795-aa14-f4f2885d8de1 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Refreshing network info cache for port 526f5c77-efba-4403-b12e-cc66f3cbd46d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 860.102231] env[62814]: DEBUG oslo_concurrency.lockutils [None req-394432b7-1628-415d-b5c1-17baa4234f90 tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "2d0b24c5-3593-4ef2-a637-d3590242ad79" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 47.149s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 860.146214] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 860.218581] env[62814]: DEBUG nova.network.neutron [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Successfully updated port: 5d635de7-0d82-4d28-a72c-35736e7d8f3b {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.463255] env[62814]: DEBUG nova.network.neutron [req-4cc5beb4-9015-4713-bdff-f44ed887a50c req-698b76f3-ad1b-4795-aa14-f4f2885d8de1 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 860.567267] env[62814]: DEBUG nova.network.neutron [req-4cc5beb4-9015-4713-bdff-f44ed887a50c req-698b76f3-ad1b-4795-aa14-f4f2885d8de1 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 860.629302] env[62814]: DEBUG nova.network.neutron [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Successfully updated port: 696f3530-9858-47e0-be99-e223f03b58fb {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.680274] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce4f8377-06b0-479b-9365-f8350fdc3b1a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.689690] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7406c60c-b2b2-47c3-bd36-5bef9f061cca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.724551] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "refresh_cache-a473cc7e-0f86-4a18-9789-0aabfff430ab" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.725804] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd023dc0-34cd-493e-abf3-2667c607a505 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.733356] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-654d8ae1-1e61-42d5-8d2c-cbb49fef2a00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.748810] env[62814]: DEBUG nova.compute.provider_tree [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.942726] env[62814]: INFO nova.compute.manager [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Rebuilding instance [ 861.000052] env[62814]: DEBUG nova.compute.manager [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 861.000052] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ecb93d-2cc7-4421-b381-5789ac96fddc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.075420] env[62814]: DEBUG oslo_concurrency.lockutils [req-4cc5beb4-9015-4713-bdff-f44ed887a50c req-698b76f3-ad1b-4795-aa14-f4f2885d8de1 service nova] Releasing lock "refresh_cache-a473cc7e-0f86-4a18-9789-0aabfff430ab" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 861.075420] env[62814]: DEBUG nova.compute.manager [req-4cc5beb4-9015-4713-bdff-f44ed887a50c req-698b76f3-ad1b-4795-aa14-f4f2885d8de1 service nova] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Received event network-vif-deleted-b6e71b20-cd42-432e-bc6d-1c16b443ed92 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 861.075420] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquired lock "refresh_cache-a473cc7e-0f86-4a18-9789-0aabfff430ab" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.075420] env[62814]: DEBUG nova.network.neutron [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.130545] env[62814]: DEBUG oslo_vmware.rw_handles [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527307af-9c2e-611e-45b8-670e7f9c168f/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 861.132039] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c59ade4-c4c5-4662-a0d2-f72951a87eb5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.136599] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquiring lock "refresh_cache-cd89e4d5-c08e-42c9-aea8-84c20714e196" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.136814] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquired lock "refresh_cache-cd89e4d5-c08e-42c9-aea8-84c20714e196" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 861.136944] env[62814]: DEBUG nova.network.neutron [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.142527] env[62814]: DEBUG oslo_vmware.rw_handles [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527307af-9c2e-611e-45b8-670e7f9c168f/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 861.143210] env[62814]: ERROR oslo_vmware.rw_handles [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527307af-9c2e-611e-45b8-670e7f9c168f/disk-0.vmdk due to incomplete transfer. [ 861.143486] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7f5ff156-a767-410c-b96d-bbe64e5dc661 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.154195] env[62814]: DEBUG oslo_vmware.rw_handles [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/527307af-9c2e-611e-45b8-670e7f9c168f/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 861.154465] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Uploaded image 577d3b83-11e9-430b-a3da-3fc2f382fd79 to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 861.156839] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 861.157403] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-71a48db1-5c5e-4723-a49d-847778c1723f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.164139] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 861.164139] env[62814]: value = "task-4293789" [ 861.164139] env[62814]: _type = "Task" [ 861.164139] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.172662] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293789, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.251930] env[62814]: DEBUG nova.scheduler.client.report [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 861.497664] env[62814]: INFO nova.compute.manager [None req-c14fdc5b-9354-42f8-8ac2-5feea74518f5 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Get console output [ 861.498022] env[62814]: WARNING nova.virt.vmwareapi.driver [None req-c14fdc5b-9354-42f8-8ac2-5feea74518f5 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] The console log is missing. Check your VSPC configuration [ 861.610765] env[62814]: DEBUG nova.network.neutron [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.673530] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293789, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.700396] env[62814]: DEBUG nova.network.neutron [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.761766] env[62814]: DEBUG oslo_concurrency.lockutils [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.204s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 861.764195] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 37.547s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 861.765783] env[62814]: INFO nova.compute.claims [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.799939] env[62814]: INFO nova.scheduler.client.report [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Deleted allocations for instance c18acec7-cf95-4cdf-aa49-32419d364534 [ 861.920008] env[62814]: DEBUG nova.network.neutron [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Updating instance_info_cache with network_info: [{"id": "526f5c77-efba-4403-b12e-cc66f3cbd46d", "address": "fa:16:3e:8d:4e:bf", "network": {"id": "ca122834-ac3d-45cb-a8aa-34c46bfcf61a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-214724823", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f87a752-ebb0-49a4-a67b-e356fa45b89b", "external-id": "nsx-vlan-transportzone-889", "segmentation_id": 889, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap526f5c77-ef", "ovs_interfaceid": "526f5c77-efba-4403-b12e-cc66f3cbd46d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5d635de7-0d82-4d28-a72c-35736e7d8f3b", "address": "fa:16:3e:8a:78:90", "network": {"id": "2dd9fea4-ea90-4ac6-9f3c-c0c3ffdfdc96", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-760127850", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.209", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da623279-b6f6-4570-8b15-a332120b8b60", "external-id": "nsx-vlan-transportzone-733", "segmentation_id": 733, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d635de7-0d", "ovs_interfaceid": "5d635de7-0d82-4d28-a72c-35736e7d8f3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.991388] env[62814]: DEBUG nova.network.neutron [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Updating instance_info_cache with network_info: [{"id": "696f3530-9858-47e0-be99-e223f03b58fb", "address": "fa:16:3e:d4:fc:a6", "network": {"id": "7efb312b-bb3e-4806-a5cd-a2122e26982e", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-651938622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f19ca23f48c40d391e7348ebf008751", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696f3530-98", "ovs_interfaceid": "696f3530-9858-47e0-be99-e223f03b58fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.013690] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 862.014353] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-326edb1c-d481-4d98-bf5b-4319dfe29078 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.021140] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 862.021140] env[62814]: value = "task-4293790" [ 862.021140] env[62814]: _type = "Task" [ 862.021140] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.029569] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293790, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.174259] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293789, 'name': Destroy_Task, 'duration_secs': 0.837848} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.174582] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Destroyed the VM [ 862.174844] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 862.175136] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-aced1509-ae9d-4256-8c99-114dbb6f7db5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.184211] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 862.184211] env[62814]: value = "task-4293791" [ 862.184211] env[62814]: _type = "Task" [ 862.184211] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.193306] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293791, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.222815] env[62814]: DEBUG nova.compute.manager [req-7c1b32a2-84bc-4270-afb1-58cb4a30e0ee req-7177aad6-6e3c-4ee8-bddd-d365f5804f51 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Received event network-changed-5d635de7-0d82-4d28-a72c-35736e7d8f3b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 862.222815] env[62814]: DEBUG nova.compute.manager [req-7c1b32a2-84bc-4270-afb1-58cb4a30e0ee req-7177aad6-6e3c-4ee8-bddd-d365f5804f51 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Refreshing instance network info cache due to event network-changed-5d635de7-0d82-4d28-a72c-35736e7d8f3b. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 862.222815] env[62814]: DEBUG oslo_concurrency.lockutils [req-7c1b32a2-84bc-4270-afb1-58cb4a30e0ee req-7177aad6-6e3c-4ee8-bddd-d365f5804f51 service nova] Acquiring lock "refresh_cache-a473cc7e-0f86-4a18-9789-0aabfff430ab" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.240457] env[62814]: DEBUG nova.compute.manager [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Received event network-vif-plugged-696f3530-9858-47e0-be99-e223f03b58fb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 862.240693] env[62814]: DEBUG oslo_concurrency.lockutils [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] Acquiring lock "cd89e4d5-c08e-42c9-aea8-84c20714e196-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.240911] env[62814]: DEBUG oslo_concurrency.lockutils [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] Lock "cd89e4d5-c08e-42c9-aea8-84c20714e196-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.241212] env[62814]: DEBUG oslo_concurrency.lockutils [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] Lock "cd89e4d5-c08e-42c9-aea8-84c20714e196-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.241495] env[62814]: DEBUG nova.compute.manager [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] No waiting events found dispatching network-vif-plugged-696f3530-9858-47e0-be99-e223f03b58fb {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 862.241584] env[62814]: WARNING nova.compute.manager [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Received unexpected event network-vif-plugged-696f3530-9858-47e0-be99-e223f03b58fb for instance with vm_state building and task_state spawning. [ 862.241742] env[62814]: DEBUG nova.compute.manager [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Received event network-changed-696f3530-9858-47e0-be99-e223f03b58fb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 862.241920] env[62814]: DEBUG nova.compute.manager [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Refreshing instance network info cache due to event network-changed-696f3530-9858-47e0-be99-e223f03b58fb. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 862.242066] env[62814]: DEBUG oslo_concurrency.lockutils [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] Acquiring lock "refresh_cache-cd89e4d5-c08e-42c9-aea8-84c20714e196" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.314789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-388aece0-7089-4053-a495-6f4c436bfd8c tempest-MultipleCreateTestJSON-872094314 tempest-MultipleCreateTestJSON-872094314-project-member] Lock "c18acec7-cf95-4cdf-aa49-32419d364534" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 49.224s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.423106] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Releasing lock "refresh_cache-a473cc7e-0f86-4a18-9789-0aabfff430ab" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 862.424523] env[62814]: DEBUG nova.compute.manager [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Instance network_info: |[{"id": "526f5c77-efba-4403-b12e-cc66f3cbd46d", "address": "fa:16:3e:8d:4e:bf", "network": {"id": "ca122834-ac3d-45cb-a8aa-34c46bfcf61a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-214724823", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f87a752-ebb0-49a4-a67b-e356fa45b89b", "external-id": "nsx-vlan-transportzone-889", "segmentation_id": 889, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap526f5c77-ef", "ovs_interfaceid": "526f5c77-efba-4403-b12e-cc66f3cbd46d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5d635de7-0d82-4d28-a72c-35736e7d8f3b", "address": "fa:16:3e:8a:78:90", "network": {"id": "2dd9fea4-ea90-4ac6-9f3c-c0c3ffdfdc96", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-760127850", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.209", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da623279-b6f6-4570-8b15-a332120b8b60", "external-id": "nsx-vlan-transportzone-733", "segmentation_id": 733, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d635de7-0d", "ovs_interfaceid": "5d635de7-0d82-4d28-a72c-35736e7d8f3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 862.424523] env[62814]: DEBUG oslo_concurrency.lockutils [req-7c1b32a2-84bc-4270-afb1-58cb4a30e0ee req-7177aad6-6e3c-4ee8-bddd-d365f5804f51 service nova] Acquired lock "refresh_cache-a473cc7e-0f86-4a18-9789-0aabfff430ab" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 862.424523] env[62814]: DEBUG nova.network.neutron [req-7c1b32a2-84bc-4270-afb1-58cb4a30e0ee req-7177aad6-6e3c-4ee8-bddd-d365f5804f51 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Refreshing network info cache for port 5d635de7-0d82-4d28-a72c-35736e7d8f3b {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.425804] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:4e:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9f87a752-ebb0-49a4-a67b-e356fa45b89b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '526f5c77-efba-4403-b12e-cc66f3cbd46d', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:78:90', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'da623279-b6f6-4570-8b15-a332120b8b60', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5d635de7-0d82-4d28-a72c-35736e7d8f3b', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.443733] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 862.446856] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.447342] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60fdf12e-b12e-4b31-a4d8-b051ec3fd01b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.472800] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.472800] env[62814]: value = "task-4293792" [ 862.472800] env[62814]: _type = "Task" [ 862.472800] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.481421] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293792, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.494917] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Releasing lock "refresh_cache-cd89e4d5-c08e-42c9-aea8-84c20714e196" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 862.495281] env[62814]: DEBUG nova.compute.manager [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Instance network_info: |[{"id": "696f3530-9858-47e0-be99-e223f03b58fb", "address": "fa:16:3e:d4:fc:a6", "network": {"id": "7efb312b-bb3e-4806-a5cd-a2122e26982e", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-651938622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f19ca23f48c40d391e7348ebf008751", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696f3530-98", "ovs_interfaceid": "696f3530-9858-47e0-be99-e223f03b58fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 862.495594] env[62814]: DEBUG oslo_concurrency.lockutils [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] Acquired lock "refresh_cache-cd89e4d5-c08e-42c9-aea8-84c20714e196" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 862.496582] env[62814]: DEBUG nova.network.neutron [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Refreshing network info cache for port 696f3530-9858-47e0-be99-e223f03b58fb {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.497383] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:fc:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd986680e-ad16-45b1-bf6d-cd2fe661679f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '696f3530-9858-47e0-be99-e223f03b58fb', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.507908] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Creating folder: Project (7f19ca23f48c40d391e7348ebf008751). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.509486] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0643125b-3874-4a17-bb1a-97f4dcae281d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.525324] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Created folder: Project (7f19ca23f48c40d391e7348ebf008751) in parent group-v845547. [ 862.525466] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Creating folder: Instances. Parent ref: group-v845705. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.529450] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-226044f0-527d-4452-bac0-6edffdecb67c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.537199] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293790, 'name': PowerOffVM_Task, 'duration_secs': 0.442101} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.537525] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 862.537785] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 862.540333] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d000b3e-53e2-4899-b063-0aa4fc8c9613 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.543256] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Created folder: Instances in parent group-v845705. [ 862.547019] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 862.547019] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.547019] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8fcb3086-14c1-4dec-9241-421ce03634b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.564099] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 862.564810] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d68af5a4-1473-409d-a650-d167f20db92b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.568841] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.568841] env[62814]: value = "task-4293795" [ 862.568841] env[62814]: _type = "Task" [ 862.568841] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.578230] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293795, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.622177] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 862.622177] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 862.622432] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleting the datastore file [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 862.631022] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-535fca3f-2d37-4d52-9c6b-4b27746bd770 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.635872] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 862.635872] env[62814]: value = "task-4293797" [ 862.635872] env[62814]: _type = "Task" [ 862.635872] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.645704] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293797, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.696424] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293791, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.897126] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "17bfe703-ff96-4cb7-8535-eed57e10d673" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.897126] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "17bfe703-ff96-4cb7-8535-eed57e10d673" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.897593] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "17bfe703-ff96-4cb7-8535-eed57e10d673-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 862.899163] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "17bfe703-ff96-4cb7-8535-eed57e10d673-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 862.899163] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "17bfe703-ff96-4cb7-8535-eed57e10d673-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.901205] env[62814]: INFO nova.compute.manager [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Terminating instance [ 862.931999] env[62814]: DEBUG nova.network.neutron [req-7c1b32a2-84bc-4270-afb1-58cb4a30e0ee req-7177aad6-6e3c-4ee8-bddd-d365f5804f51 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Updated VIF entry in instance network info cache for port 5d635de7-0d82-4d28-a72c-35736e7d8f3b. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 862.932389] env[62814]: DEBUG nova.network.neutron [req-7c1b32a2-84bc-4270-afb1-58cb4a30e0ee req-7177aad6-6e3c-4ee8-bddd-d365f5804f51 service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Updating instance_info_cache with network_info: [{"id": "526f5c77-efba-4403-b12e-cc66f3cbd46d", "address": "fa:16:3e:8d:4e:bf", "network": {"id": "ca122834-ac3d-45cb-a8aa-34c46bfcf61a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-214724823", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f87a752-ebb0-49a4-a67b-e356fa45b89b", "external-id": "nsx-vlan-transportzone-889", "segmentation_id": 889, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap526f5c77-ef", "ovs_interfaceid": "526f5c77-efba-4403-b12e-cc66f3cbd46d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "5d635de7-0d82-4d28-a72c-35736e7d8f3b", "address": "fa:16:3e:8a:78:90", "network": {"id": "2dd9fea4-ea90-4ac6-9f3c-c0c3ffdfdc96", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-760127850", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.209", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "da623279-b6f6-4570-8b15-a332120b8b60", "external-id": "nsx-vlan-transportzone-733", "segmentation_id": 733, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d635de7-0d", "ovs_interfaceid": "5d635de7-0d82-4d28-a72c-35736e7d8f3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.984490] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293792, 'name': CreateVM_Task, 'duration_secs': 0.380132} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.987065] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.988332] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.988502] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 862.988827] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 862.989117] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b68a62a4-c5a8-4d4e-a8ad-e91d0bd0a2d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.994360] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 862.994360] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b9c015-fa6c-7081-b7e7-38da10ea914c" [ 862.994360] env[62814]: _type = "Task" [ 862.994360] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.003033] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b9c015-fa6c-7081-b7e7-38da10ea914c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.083616] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293795, 'name': CreateVM_Task, 'duration_secs': 0.39256} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.083802] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 863.084499] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.146609] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293797, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228529} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.149485] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.149838] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 863.150143] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 863.197940] env[62814]: DEBUG oslo_vmware.api [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293791, 'name': RemoveSnapshot_Task, 'duration_secs': 0.901773} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.201826] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 863.201955] env[62814]: INFO nova.compute.manager [None req-6d8049c4-40d5-4065-a651-2e60448b0566 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Took 16.20 seconds to snapshot the instance on the hypervisor. [ 863.375323] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa80d038-038e-45f8-96cd-e9cce5a8c2e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.383476] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7604c70c-256a-4871-9ab8-6f2ae36e856a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.417201] env[62814]: DEBUG nova.compute.manager [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 863.417562] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 863.418512] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b87dca9a-8678-46ea-8668-9d4da2da53d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.421992] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22799b56-0805-46b4-a2af-fe1207ce5ef5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.432692] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82afdd45-b967-4f62-a9e9-e15b9c0bb9c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.437380] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.437764] env[62814]: DEBUG oslo_concurrency.lockutils [req-7c1b32a2-84bc-4270-afb1-58cb4a30e0ee req-7177aad6-6e3c-4ee8-bddd-d365f5804f51 service nova] Releasing lock "refresh_cache-a473cc7e-0f86-4a18-9789-0aabfff430ab" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.438102] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9704723f-274d-44dc-8e45-690dd4eceec3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.445149] env[62814]: DEBUG nova.network.neutron [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Updated VIF entry in instance network info cache for port 696f3530-9858-47e0-be99-e223f03b58fb. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.445316] env[62814]: DEBUG nova.network.neutron [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Updating instance_info_cache with network_info: [{"id": "696f3530-9858-47e0-be99-e223f03b58fb", "address": "fa:16:3e:d4:fc:a6", "network": {"id": "7efb312b-bb3e-4806-a5cd-a2122e26982e", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-651938622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f19ca23f48c40d391e7348ebf008751", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696f3530-98", "ovs_interfaceid": "696f3530-9858-47e0-be99-e223f03b58fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.460737] env[62814]: DEBUG nova.compute.provider_tree [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.462528] env[62814]: DEBUG oslo_vmware.api [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 863.462528] env[62814]: value = "task-4293798" [ 863.462528] env[62814]: _type = "Task" [ 863.462528] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.472592] env[62814]: DEBUG oslo_vmware.api [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293798, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.504765] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b9c015-fa6c-7081-b7e7-38da10ea914c, 'name': SearchDatastore_Task, 'duration_secs': 0.011896} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.507550] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.507550] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.507550] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.507550] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.507550] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.507550] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 863.507550] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 863.507550] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b53fa0ab-89a1-43a9-89b6-13bd5ce5cee1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.510134] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff984327-6963-4a38-91d8-101d069c37e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.516256] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for the task: (returnval){ [ 863.516256] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d159f3-9dac-da54-5e9b-0b61ad2c2f1b" [ 863.516256] env[62814]: _type = "Task" [ 863.516256] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.523045] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.526037] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.526037] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76cac4b5-a999-459a-8cb8-a55b70212c56 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.529983] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d159f3-9dac-da54-5e9b-0b61ad2c2f1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.533165] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 863.533165] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb20d6-91f1-b709-396b-f23bf6e9e085" [ 863.533165] env[62814]: _type = "Task" [ 863.533165] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.543197] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb20d6-91f1-b709-396b-f23bf6e9e085, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.848384] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "af6ef867-dc9a-4db6-8582-a6198e5caa77" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 863.848953] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "af6ef867-dc9a-4db6-8582-a6198e5caa77" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 863.964903] env[62814]: DEBUG nova.scheduler.client.report [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 863.968468] env[62814]: DEBUG oslo_concurrency.lockutils [req-0fd6f017-f41a-4f6e-8b35-2c3863a77ea8 req-d55645a6-691c-4232-8974-43d9b097c1f4 service nova] Releasing lock "refresh_cache-cd89e4d5-c08e-42c9-aea8-84c20714e196" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 863.979321] env[62814]: DEBUG oslo_vmware.api [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293798, 'name': PowerOffVM_Task, 'duration_secs': 0.329227} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.979639] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.979826] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 863.980669] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c19f2da-82e7-4fa1-89b6-b63b8e27de01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.027584] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d159f3-9dac-da54-5e9b-0b61ad2c2f1b, 'name': SearchDatastore_Task, 'duration_secs': 0.012933} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.027937] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.028197] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.028806] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.040884] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.041041] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.041224] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Deleting the datastore file [datastore2] 17bfe703-ff96-4cb7-8535-eed57e10d673 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.042054] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0675f2cf-2c9d-4a07-9991-0ea420ea7b66 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.048994] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb20d6-91f1-b709-396b-f23bf6e9e085, 'name': SearchDatastore_Task, 'duration_secs': 0.011421} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.049298] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdd879db-1256-4318-a093-b8081a9a2431 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.052938] env[62814]: DEBUG oslo_vmware.api [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 864.052938] env[62814]: value = "task-4293800" [ 864.052938] env[62814]: _type = "Task" [ 864.052938] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.058484] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 864.058484] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522a2cf3-7742-d20a-7f47-ed68b40a1957" [ 864.058484] env[62814]: _type = "Task" [ 864.058484] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.065715] env[62814]: DEBUG oslo_vmware.api [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293800, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.070574] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522a2cf3-7742-d20a-7f47-ed68b40a1957, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.196543] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 864.196846] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.197097] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 864.197373] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.197837] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 864.197837] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 864.197983] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 864.198223] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 864.198348] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 864.198790] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 864.199045] env[62814]: DEBUG nova.virt.hardware [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 864.200970] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3095f2aa-381d-4bc4-806f-38d879887f86 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.212061] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd9c48e-2891-4d97-9345-93bed7790a38 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.230082] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:cb:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b7654ff-51f0-4a51-9b69-04d3352ddbee', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.237784] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 864.238145] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.238508] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1ec02c46-2d4e-4bf0-872e-f3a1d2bad6ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.260502] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.260502] env[62814]: value = "task-4293801" [ 864.260502] env[62814]: _type = "Task" [ 864.260502] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.266487] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293801, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.352293] env[62814]: DEBUG nova.compute.manager [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 864.475381] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.711s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 864.475962] env[62814]: DEBUG nova.compute.manager [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 864.478944] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 39.621s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 864.566000] env[62814]: DEBUG oslo_vmware.api [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4293800, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.433254} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.568736] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.568736] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 864.568736] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.568736] env[62814]: INFO nova.compute.manager [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Took 1.15 seconds to destroy the instance on the hypervisor. [ 864.568736] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 864.568736] env[62814]: DEBUG nova.compute.manager [-] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 864.568736] env[62814]: DEBUG nova.network.neutron [-] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 864.573615] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522a2cf3-7742-d20a-7f47-ed68b40a1957, 'name': SearchDatastore_Task, 'duration_secs': 0.03847} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.574217] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 864.574471] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a473cc7e-0f86-4a18-9789-0aabfff430ab/a473cc7e-0f86-4a18-9789-0aabfff430ab.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.574746] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 864.574932] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.575161] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f60c5c1c-4159-4d39-83ce-78ce035f650c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.577651] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab692131-8441-4dca-bb4e-cac8b3210001 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.585180] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 864.585180] env[62814]: value = "task-4293802" [ 864.585180] env[62814]: _type = "Task" [ 864.585180] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.590051] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 864.590051] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 864.590412] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a89548ba-a2cf-49cc-b7ba-322efab1293a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.595904] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293802, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.599591] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for the task: (returnval){ [ 864.599591] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5285c7a8-0fc4-954a-742a-82a1f2144cd4" [ 864.599591] env[62814]: _type = "Task" [ 864.599591] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.607845] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5285c7a8-0fc4-954a-742a-82a1f2144cd4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.770345] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293801, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.881310] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.991026] env[62814]: INFO nova.compute.claims [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.993972] env[62814]: DEBUG nova.compute.utils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 864.997104] env[62814]: DEBUG nova.compute.manager [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 864.997559] env[62814]: DEBUG nova.network.neutron [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 865.096412] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293802, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.109189] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5285c7a8-0fc4-954a-742a-82a1f2144cd4, 'name': SearchDatastore_Task, 'duration_secs': 0.013371} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.110124] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8471fab1-f9b7-4592-9902-d3b4d9e197de {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.116084] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for the task: (returnval){ [ 865.116084] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52eaddb0-8013-49c9-be8b-f69142d9c8a0" [ 865.116084] env[62814]: _type = "Task" [ 865.116084] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.125595] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52eaddb0-8013-49c9-be8b-f69142d9c8a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.138872] env[62814]: DEBUG nova.policy [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '093788d5d63e4c6a84750ffaf1e31455', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d51711cd907404b8d0a48d904b599c4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 865.141293] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "7350d352-9336-40b8-81a6-0a4795d9f8dd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 865.141625] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "7350d352-9336-40b8-81a6-0a4795d9f8dd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 865.272918] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293801, 'name': CreateVM_Task, 'duration_secs': 0.512982} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.272918] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.276766] env[62814]: DEBUG oslo_concurrency.lockutils [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.277179] env[62814]: DEBUG oslo_concurrency.lockutils [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.277282] env[62814]: DEBUG oslo_concurrency.lockutils [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 865.277540] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e84db6cc-e10e-4278-933d-1b9ed4d887ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.283026] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 865.283026] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529d42dc-3575-1492-fa93-e9a0947c2811" [ 865.283026] env[62814]: _type = "Task" [ 865.283026] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.298604] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529d42dc-3575-1492-fa93-e9a0947c2811, 'name': SearchDatastore_Task, 'duration_secs': 0.011078} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.299122] env[62814]: DEBUG oslo_concurrency.lockutils [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.299122] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.299342] env[62814]: DEBUG oslo_concurrency.lockutils [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.449161] env[62814]: DEBUG nova.compute.manager [req-bca98cd0-d82e-4fb8-bfb6-c10e317ac112 req-ef66052f-79ba-4ffb-807a-b87a2131e1b4 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Received event network-vif-deleted-6e1ae96b-a9fd-4e11-9803-191a35a77cb2 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 865.449617] env[62814]: INFO nova.compute.manager [req-bca98cd0-d82e-4fb8-bfb6-c10e317ac112 req-ef66052f-79ba-4ffb-807a-b87a2131e1b4 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Neutron deleted interface 6e1ae96b-a9fd-4e11-9803-191a35a77cb2; detaching it from the instance and deleting it from the info cache [ 865.449617] env[62814]: DEBUG nova.network.neutron [req-bca98cd0-d82e-4fb8-bfb6-c10e317ac112 req-ef66052f-79ba-4ffb-807a-b87a2131e1b4 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.498022] env[62814]: DEBUG nova.compute.utils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 865.501706] env[62814]: INFO nova.compute.resource_tracker [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating resource usage from migration 8d6e5dee-326f-4a97-a651-a24c301bccc8 [ 865.576022] env[62814]: DEBUG nova.network.neutron [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Successfully created port: 5ec47315-906b-4ad9-a8c8-87622cb99c27 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.597232] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293802, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.675119} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.601098] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a473cc7e-0f86-4a18-9789-0aabfff430ab/a473cc7e-0f86-4a18-9789-0aabfff430ab.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 865.601631] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.603214] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-99f203b2-1c33-4e93-be57-5e2ac580063c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.613075] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 865.613075] env[62814]: value = "task-4293803" [ 865.613075] env[62814]: _type = "Task" [ 865.613075] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.627633] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293803, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.633276] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52eaddb0-8013-49c9-be8b-f69142d9c8a0, 'name': SearchDatastore_Task, 'duration_secs': 0.062495} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.633904] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 865.634310] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] cd89e4d5-c08e-42c9-aea8-84c20714e196/cd89e4d5-c08e-42c9-aea8-84c20714e196.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 865.634653] env[62814]: DEBUG oslo_concurrency.lockutils [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.634907] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.635143] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d155e8f1-28a6-4b04-b7aa-f2b96fe67dbe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.637381] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05358326-6d4e-4e86-a983-c5c4822b702e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.648165] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for the task: (returnval){ [ 865.648165] env[62814]: value = "task-4293804" [ 865.648165] env[62814]: _type = "Task" [ 865.648165] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.652532] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.652863] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.654822] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55c88ab5-6d1e-49cb-ad76-b2350b2177bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.660473] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.668018] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 865.668018] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bf8ee3-0f73-740c-8db0-0a0a6384e3a2" [ 865.668018] env[62814]: _type = "Task" [ 865.668018] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.674715] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bf8ee3-0f73-740c-8db0-0a0a6384e3a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.930495] env[62814]: DEBUG nova.network.neutron [-] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.952227] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-930fc83d-2212-4e5a-b9fa-b86b76704519 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.961698] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdf4fd91-824b-4980-89f3-322d2373156d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.007096] env[62814]: DEBUG nova.compute.manager [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 866.010608] env[62814]: DEBUG nova.compute.manager [req-bca98cd0-d82e-4fb8-bfb6-c10e317ac112 req-ef66052f-79ba-4ffb-807a-b87a2131e1b4 service nova] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Detach interface failed, port_id=6e1ae96b-a9fd-4e11-9803-191a35a77cb2, reason: Instance 17bfe703-ff96-4cb7-8535-eed57e10d673 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 866.127673] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293803, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085369} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.128033] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.128874] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45015d41-0b1b-4f33-9057-1d1ac45d2996 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.154605] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] a473cc7e-0f86-4a18-9789-0aabfff430ab/a473cc7e-0f86-4a18-9789-0aabfff430ab.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.157677] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efd99bf3-a8aa-4396-934c-2f236b2cef51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.182595] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293804, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.188860] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 866.188860] env[62814]: value = "task-4293805" [ 866.188860] env[62814]: _type = "Task" [ 866.188860] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.189097] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bf8ee3-0f73-740c-8db0-0a0a6384e3a2, 'name': SearchDatastore_Task, 'duration_secs': 0.01811} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.192545] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d8c7281-1b89-4092-883c-ee15b8b44f86 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.199218] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293805, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.201641] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b31d39-d662-4bce-a0f7-e1f856841942 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.204031] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 866.204031] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52abbdfd-ab5e-e329-75c4-8b062037e9bf" [ 866.204031] env[62814]: _type = "Task" [ 866.204031] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.210171] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c111123b-196c-481e-ad22-07f94f190c21 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.216188] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52abbdfd-ab5e-e329-75c4-8b062037e9bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.244924] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a81f67cf-d78e-4464-9a68-816a8c425092 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.252681] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ceee770-f8f9-4338-8b85-ca33c7e896d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.267072] env[62814]: DEBUG nova.compute.provider_tree [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.433984] env[62814]: INFO nova.compute.manager [-] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Took 1.87 seconds to deallocate network for instance. [ 866.667389] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293804, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.935628} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.667656] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] cd89e4d5-c08e-42c9-aea8-84c20714e196/cd89e4d5-c08e-42c9-aea8-84c20714e196.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.667871] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.668145] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a2a9f49-4007-434e-b4b9-fb58c44d9649 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.676029] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for the task: (returnval){ [ 866.676029] env[62814]: value = "task-4293806" [ 866.676029] env[62814]: _type = "Task" [ 866.676029] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.683314] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293806, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.699356] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293805, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.713214] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52abbdfd-ab5e-e329-75c4-8b062037e9bf, 'name': SearchDatastore_Task, 'duration_secs': 0.019458} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.713526] env[62814]: DEBUG oslo_concurrency.lockutils [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.713817] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.714115] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f85d464e-7cea-40de-b6b3-37fef40b2d87 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.721174] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 866.721174] env[62814]: value = "task-4293807" [ 866.721174] env[62814]: _type = "Task" [ 866.721174] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.728748] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293807, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.770085] env[62814]: DEBUG nova.scheduler.client.report [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 866.941934] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.021707] env[62814]: DEBUG nova.compute.manager [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 867.050647] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:34:23Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=1,extra_specs={hw_rng:allowed='True'},flavorid='823943732',id=18,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_1-1368852159',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 867.050949] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 867.051133] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 867.051355] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 867.051505] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 867.051645] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 867.051882] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 867.052067] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 867.052275] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 867.052489] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 867.052723] env[62814]: DEBUG nova.virt.hardware [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 867.053966] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f5c356-570a-4279-9144-374b3c003007 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.062190] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7063550-0803-41e7-a053-88b67e415fc0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.183943] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293806, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069864} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.184247] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.185073] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9405f005-e38d-475d-8704-9b8d7ba6df2a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.209574] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] cd89e4d5-c08e-42c9-aea8-84c20714e196/cd89e4d5-c08e-42c9-aea8-84c20714e196.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.212727] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-087485a9-6a89-49aa-a378-1b760b2532e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.228909] env[62814]: DEBUG nova.compute.manager [req-65a5608f-1df0-421e-890c-ba8d79c9d788 req-b0829cf8-4f29-4ae7-ad9c-64308de8502e service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Received event network-vif-plugged-5ec47315-906b-4ad9-a8c8-87622cb99c27 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 867.229135] env[62814]: DEBUG oslo_concurrency.lockutils [req-65a5608f-1df0-421e-890c-ba8d79c9d788 req-b0829cf8-4f29-4ae7-ad9c-64308de8502e service nova] Acquiring lock "9be62576-5a05-473f-befd-b33f5fde9185-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 867.229343] env[62814]: DEBUG oslo_concurrency.lockutils [req-65a5608f-1df0-421e-890c-ba8d79c9d788 req-b0829cf8-4f29-4ae7-ad9c-64308de8502e service nova] Lock "9be62576-5a05-473f-befd-b33f5fde9185-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.229508] env[62814]: DEBUG oslo_concurrency.lockutils [req-65a5608f-1df0-421e-890c-ba8d79c9d788 req-b0829cf8-4f29-4ae7-ad9c-64308de8502e service nova] Lock "9be62576-5a05-473f-befd-b33f5fde9185-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.229674] env[62814]: DEBUG nova.compute.manager [req-65a5608f-1df0-421e-890c-ba8d79c9d788 req-b0829cf8-4f29-4ae7-ad9c-64308de8502e service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] No waiting events found dispatching network-vif-plugged-5ec47315-906b-4ad9-a8c8-87622cb99c27 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 867.229831] env[62814]: WARNING nova.compute.manager [req-65a5608f-1df0-421e-890c-ba8d79c9d788 req-b0829cf8-4f29-4ae7-ad9c-64308de8502e service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Received unexpected event network-vif-plugged-5ec47315-906b-4ad9-a8c8-87622cb99c27 for instance with vm_state building and task_state spawning. [ 867.238086] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293805, 'name': ReconfigVM_Task, 'duration_secs': 0.553505} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.238777] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Reconfigured VM instance instance-00000038 to attach disk [datastore2] a473cc7e-0f86-4a18-9789-0aabfff430ab/a473cc7e-0f86-4a18-9789-0aabfff430ab.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.239492] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-600d2cf0-0042-4422-8f93-708bb9dfd5c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.245768] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for the task: (returnval){ [ 867.245768] env[62814]: value = "task-4293808" [ 867.245768] env[62814]: _type = "Task" [ 867.245768] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.246261] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293807, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457835} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.246736] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 867.247036] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.250084] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c220e73-eff2-441b-960b-83f41c002141 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.253428] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 867.253428] env[62814]: value = "task-4293809" [ 867.253428] env[62814]: _type = "Task" [ 867.253428] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.260311] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293808, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.261825] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 867.261825] env[62814]: value = "task-4293810" [ 867.261825] env[62814]: _type = "Task" [ 867.261825] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.268017] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293809, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.273381] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293810, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.275235] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.796s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 867.275510] env[62814]: INFO nova.compute.manager [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Migrating [ 867.282276] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 41.296s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 867.282618] env[62814]: DEBUG nova.objects.instance [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lazy-loading 'resources' on Instance uuid 5087b202-9bba-4489-823b-5d93cbf116e2 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.297734] env[62814]: DEBUG nova.network.neutron [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Successfully updated port: 5ec47315-906b-4ad9-a8c8-87622cb99c27 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 867.757137] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293808, 'name': ReconfigVM_Task, 'duration_secs': 0.285748} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.760203] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Reconfigured VM instance instance-00000039 to attach disk [datastore2] cd89e4d5-c08e-42c9-aea8-84c20714e196/cd89e4d5-c08e-42c9-aea8-84c20714e196.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.760847] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0520f116-7fd1-4cf4-89fb-90b933d6100e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.768548] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293809, 'name': Rename_Task, 'duration_secs': 0.149326} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.769118] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 867.770187] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57f5adbe-349e-4d42-84a3-a1ad332bd6ee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.771576] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for the task: (returnval){ [ 867.771576] env[62814]: value = "task-4293811" [ 867.771576] env[62814]: _type = "Task" [ 867.771576] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.774297] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293810, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072107} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.777227] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.778737] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943b44a9-5c5f-4048-b407-6c7ca749adbd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.781183] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 867.781183] env[62814]: value = "task-4293812" [ 867.781183] env[62814]: _type = "Task" [ 867.781183] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.791237] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293811, 'name': Rename_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.804095] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.804316] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 867.804490] env[62814]: DEBUG nova.network.neutron [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 867.806145] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "refresh_cache-9be62576-5a05-473f-befd-b33f5fde9185" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.806278] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquired lock "refresh_cache-9be62576-5a05-473f-befd-b33f5fde9185" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 867.806421] env[62814]: DEBUG nova.network.neutron [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 867.820076] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.831613] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e14ec4f7-e07c-4ba3-a487-15898a897cad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.847719] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293812, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.853149] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 867.853149] env[62814]: value = "task-4293813" [ 867.853149] env[62814]: _type = "Task" [ 867.853149] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.863875] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293813, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.880429] env[62814]: DEBUG nova.network.neutron [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 868.098798] env[62814]: DEBUG nova.network.neutron [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Updating instance_info_cache with network_info: [{"id": "5ec47315-906b-4ad9-a8c8-87622cb99c27", "address": "fa:16:3e:5f:99:cd", "network": {"id": "b4436b99-a4a3-46c9-918a-4e6f6f6bb55e", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-2020231604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d51711cd907404b8d0a48d904b599c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98d96b75-ac36-499a-adc2-130c8c1d55ca", "external-id": "nsx-vlan-transportzone-564", "segmentation_id": 564, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ec47315-90", "ovs_interfaceid": "5ec47315-906b-4ad9-a8c8-87622cb99c27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.291063] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293811, 'name': Rename_Task, 'duration_secs': 0.142924} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.291063] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.291063] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-647d9ccc-d69a-41f1-b9db-3c2aab1e3783 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.297943] env[62814]: DEBUG oslo_vmware.api [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293812, 'name': PowerOnVM_Task, 'duration_secs': 0.491659} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.298707] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.300046] env[62814]: INFO nova.compute.manager [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Took 11.63 seconds to spawn the instance on the hypervisor. [ 868.300046] env[62814]: DEBUG nova.compute.manager [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 868.300046] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1dbd1b-efe7-446b-86c5-5118dc0209b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.303454] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for the task: (returnval){ [ 868.303454] env[62814]: value = "task-4293814" [ 868.303454] env[62814]: _type = "Task" [ 868.303454] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.319609] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293814, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.365175] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293813, 'name': ReconfigVM_Task, 'duration_secs': 0.27031} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.369650] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Reconfigured VM instance instance-00000016 to attach disk [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.371248] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-05358784-6e19-439d-b846-15f70ab876f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.374831] env[62814]: DEBUG nova.network.neutron [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance_info_cache with network_info: [{"id": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "address": "fa:16:3e:fe:5b:6c", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38976c-88", "ovs_interfaceid": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.382909] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 868.382909] env[62814]: value = "task-4293815" [ 868.382909] env[62814]: _type = "Task" [ 868.382909] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.384676] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f70c130-bb13-4ce5-bffa-cb27fc0cdb20 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.398630] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52377c00-9c83-4c66-9214-316fc14a301b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.402139] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293815, 'name': Rename_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.433287] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1980c291-4ea3-474a-900e-084ad862857c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.441094] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32db192e-68bf-460d-82e5-51d885ed6a70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.455417] env[62814]: DEBUG nova.compute.provider_tree [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.607208] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Releasing lock "refresh_cache-9be62576-5a05-473f-befd-b33f5fde9185" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 868.607589] env[62814]: DEBUG nova.compute.manager [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Instance network_info: |[{"id": "5ec47315-906b-4ad9-a8c8-87622cb99c27", "address": "fa:16:3e:5f:99:cd", "network": {"id": "b4436b99-a4a3-46c9-918a-4e6f6f6bb55e", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-2020231604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d51711cd907404b8d0a48d904b599c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98d96b75-ac36-499a-adc2-130c8c1d55ca", "external-id": "nsx-vlan-transportzone-564", "segmentation_id": 564, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ec47315-90", "ovs_interfaceid": "5ec47315-906b-4ad9-a8c8-87622cb99c27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 868.608057] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:99:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98d96b75-ac36-499a-adc2-130c8c1d55ca', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ec47315-906b-4ad9-a8c8-87622cb99c27', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 868.615663] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 868.615878] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 868.616112] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b69412e-26d0-4a67-bdcc-5af801826a0a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.638815] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 868.638815] env[62814]: value = "task-4293816" [ 868.638815] env[62814]: _type = "Task" [ 868.638815] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.646839] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293816, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.813306] env[62814]: DEBUG oslo_vmware.api [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293814, 'name': PowerOnVM_Task, 'duration_secs': 0.506358} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.813611] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 868.813782] env[62814]: INFO nova.compute.manager [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Took 9.41 seconds to spawn the instance on the hypervisor. [ 868.813960] env[62814]: DEBUG nova.compute.manager [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 868.814792] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1d74b3-ace6-45dc-af8f-21522905081e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.827612] env[62814]: INFO nova.compute.manager [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Took 55.01 seconds to build instance. [ 868.877835] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 868.894829] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293815, 'name': Rename_Task, 'duration_secs': 0.199673} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.895724] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.896018] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6771c2f3-abf1-4732-a685-b33e11d421ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.902956] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 868.902956] env[62814]: value = "task-4293817" [ 868.902956] env[62814]: _type = "Task" [ 868.902956] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.911768] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293817, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.958964] env[62814]: DEBUG nova.scheduler.client.report [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 869.148495] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293816, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.275261] env[62814]: DEBUG nova.compute.manager [req-3523545b-229d-4e42-bcc9-6754a4a16fb2 req-08f2df11-33d9-4411-8e4f-cc6fecccb63f service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Received event network-changed-5ec47315-906b-4ad9-a8c8-87622cb99c27 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 869.275459] env[62814]: DEBUG nova.compute.manager [req-3523545b-229d-4e42-bcc9-6754a4a16fb2 req-08f2df11-33d9-4411-8e4f-cc6fecccb63f service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Refreshing instance network info cache due to event network-changed-5ec47315-906b-4ad9-a8c8-87622cb99c27. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 869.275678] env[62814]: DEBUG oslo_concurrency.lockutils [req-3523545b-229d-4e42-bcc9-6754a4a16fb2 req-08f2df11-33d9-4411-8e4f-cc6fecccb63f service nova] Acquiring lock "refresh_cache-9be62576-5a05-473f-befd-b33f5fde9185" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.275824] env[62814]: DEBUG oslo_concurrency.lockutils [req-3523545b-229d-4e42-bcc9-6754a4a16fb2 req-08f2df11-33d9-4411-8e4f-cc6fecccb63f service nova] Acquired lock "refresh_cache-9be62576-5a05-473f-befd-b33f5fde9185" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 869.275987] env[62814]: DEBUG nova.network.neutron [req-3523545b-229d-4e42-bcc9-6754a4a16fb2 req-08f2df11-33d9-4411-8e4f-cc6fecccb63f service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Refreshing network info cache for port 5ec47315-906b-4ad9-a8c8-87622cb99c27 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 869.333861] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3cc242b9-f9d8-4b7e-900b-8ba68360f116 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "a473cc7e-0f86-4a18-9789-0aabfff430ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.256s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.337420] env[62814]: INFO nova.compute.manager [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Took 55.32 seconds to build instance. [ 869.413632] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293817, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.463966] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.182s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.466381] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 38.347s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.468501] env[62814]: INFO nova.compute.claims [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.636186] env[62814]: INFO nova.scheduler.client.report [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Deleted allocations for instance 5087b202-9bba-4489-823b-5d93cbf116e2 [ 869.649218] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293816, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.838877] env[62814]: DEBUG nova.compute.manager [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 869.841550] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ad573b3-23b9-4955-bf1f-8ad81c95de47 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lock "cd89e4d5-c08e-42c9-aea8-84c20714e196" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.783s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.866999] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "a473cc7e-0f86-4a18-9789-0aabfff430ab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.866999] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "a473cc7e-0f86-4a18-9789-0aabfff430ab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.866999] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "a473cc7e-0f86-4a18-9789-0aabfff430ab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 869.866999] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "a473cc7e-0f86-4a18-9789-0aabfff430ab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 869.866999] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "a473cc7e-0f86-4a18-9789-0aabfff430ab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 869.867810] env[62814]: INFO nova.compute.manager [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Terminating instance [ 869.913245] env[62814]: DEBUG oslo_vmware.api [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293817, 'name': PowerOnVM_Task, 'duration_secs': 0.993933} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.913715] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.913902] env[62814]: DEBUG nova.compute.manager [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 869.914884] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361e54f0-6ae3-4991-ad6b-e0d72caff01e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.078941] env[62814]: DEBUG nova.network.neutron [req-3523545b-229d-4e42-bcc9-6754a4a16fb2 req-08f2df11-33d9-4411-8e4f-cc6fecccb63f service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Updated VIF entry in instance network info cache for port 5ec47315-906b-4ad9-a8c8-87622cb99c27. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 870.079392] env[62814]: DEBUG nova.network.neutron [req-3523545b-229d-4e42-bcc9-6754a4a16fb2 req-08f2df11-33d9-4411-8e4f-cc6fecccb63f service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Updating instance_info_cache with network_info: [{"id": "5ec47315-906b-4ad9-a8c8-87622cb99c27", "address": "fa:16:3e:5f:99:cd", "network": {"id": "b4436b99-a4a3-46c9-918a-4e6f6f6bb55e", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-2020231604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d51711cd907404b8d0a48d904b599c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98d96b75-ac36-499a-adc2-130c8c1d55ca", "external-id": "nsx-vlan-transportzone-564", "segmentation_id": 564, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ec47315-90", "ovs_interfaceid": "5ec47315-906b-4ad9-a8c8-87622cb99c27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.147112] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ce5e0c25-69a5-4d3b-8dfa-07ef1be62527 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "5087b202-9bba-4489-823b-5d93cbf116e2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 48.459s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 870.155583] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293816, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.360820] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 870.372721] env[62814]: DEBUG nova.compute.manager [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 870.372948] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 870.373839] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf6c4a3-ab68-4505-ac3a-a048cdb9dcff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.381485] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 870.381726] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4703ad12-346d-460a-8943-a27529ede788 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.387862] env[62814]: DEBUG oslo_vmware.api [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 870.387862] env[62814]: value = "task-4293818" [ 870.387862] env[62814]: _type = "Task" [ 870.387862] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.394729] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf1568a0-5cae-487b-9cf2-0e260500a1ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.401039] env[62814]: DEBUG oslo_vmware.api [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293818, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.417067] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance '9cd89da4-d190-4bfc-81e1-da15b98d54ba' progress to 0 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 870.431504] env[62814]: DEBUG oslo_concurrency.lockutils [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 870.583125] env[62814]: DEBUG oslo_concurrency.lockutils [req-3523545b-229d-4e42-bcc9-6754a4a16fb2 req-08f2df11-33d9-4411-8e4f-cc6fecccb63f service nova] Releasing lock "refresh_cache-9be62576-5a05-473f-befd-b33f5fde9185" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 870.653164] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293816, 'name': CreateVM_Task, 'duration_secs': 1.644567} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.653372] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.654175] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.654351] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 870.654718] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 870.654963] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27844259-3664-4ab0-aaf2-e855b0b11c58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.660194] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 870.660194] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52558474-0881-61e7-7ac7-98cc8506c0d5" [ 870.660194] env[62814]: _type = "Task" [ 870.660194] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.671961] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52558474-0881-61e7-7ac7-98cc8506c0d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.897307] env[62814]: DEBUG oslo_vmware.api [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293818, 'name': PowerOffVM_Task, 'duration_secs': 0.328864} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.897638] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.897748] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 870.898771] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb9e3d6d-60d0-4540-9a7e-e4313c44fea4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.923258] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 870.926238] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f00e396c-3fb2-4b82-96b0-03498b382cb8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.934043] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 870.934043] env[62814]: value = "task-4293820" [ 870.934043] env[62814]: _type = "Task" [ 870.934043] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.941887] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293820, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.005271] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 871.005721] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 871.005721] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Deleting the datastore file [datastore2] a473cc7e-0f86-4a18-9789-0aabfff430ab {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 871.008757] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44204542-795b-44d4-9cc1-8b3e9d498911 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.015704] env[62814]: DEBUG oslo_vmware.api [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for the task: (returnval){ [ 871.015704] env[62814]: value = "task-4293821" [ 871.015704] env[62814]: _type = "Task" [ 871.015704] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.026241] env[62814]: DEBUG oslo_vmware.api [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293821, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.042270] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2270b281-81bc-4d13-a4a0-3e79d50897bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.049656] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb544a8-40e0-47de-85b3-a8bf58c934c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.090100] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4353457-149d-4adb-9a42-592cb45a7f01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.108115] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-766461c6-fdf1-4001-b007-04ff472089ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.123362] env[62814]: DEBUG nova.compute.provider_tree [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.173676] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52558474-0881-61e7-7ac7-98cc8506c0d5, 'name': SearchDatastore_Task, 'duration_secs': 0.011975} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.174335] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 871.174576] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.175188] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.175188] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 871.175188] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.175479] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-200a012c-485c-4af1-89e6-edb2e0cad0f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.192092] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.192299] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.193067] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89fb08ea-79b0-4d91-8b61-e097e1bdabf8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.198326] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 871.198326] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524436ec-587d-3f21-1c06-28e17114013e" [ 871.198326] env[62814]: _type = "Task" [ 871.198326] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.206026] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524436ec-587d-3f21-1c06-28e17114013e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.301981] env[62814]: DEBUG nova.compute.manager [req-a32a7a9a-a83a-4b3e-978d-fa9aa9138269 req-c2fba204-e3ff-4d79-a226-ce56804983ef service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Received event network-changed-696f3530-9858-47e0-be99-e223f03b58fb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 871.302187] env[62814]: DEBUG nova.compute.manager [req-a32a7a9a-a83a-4b3e-978d-fa9aa9138269 req-c2fba204-e3ff-4d79-a226-ce56804983ef service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Refreshing instance network info cache due to event network-changed-696f3530-9858-47e0-be99-e223f03b58fb. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 871.302400] env[62814]: DEBUG oslo_concurrency.lockutils [req-a32a7a9a-a83a-4b3e-978d-fa9aa9138269 req-c2fba204-e3ff-4d79-a226-ce56804983ef service nova] Acquiring lock "refresh_cache-cd89e4d5-c08e-42c9-aea8-84c20714e196" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.302541] env[62814]: DEBUG oslo_concurrency.lockutils [req-a32a7a9a-a83a-4b3e-978d-fa9aa9138269 req-c2fba204-e3ff-4d79-a226-ce56804983ef service nova] Acquired lock "refresh_cache-cd89e4d5-c08e-42c9-aea8-84c20714e196" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 871.302697] env[62814]: DEBUG nova.network.neutron [req-a32a7a9a-a83a-4b3e-978d-fa9aa9138269 req-c2fba204-e3ff-4d79-a226-ce56804983ef service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Refreshing network info cache for port 696f3530-9858-47e0-be99-e223f03b58fb {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.443132] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293820, 'name': PowerOffVM_Task, 'duration_secs': 0.186213} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.443320] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 871.443507] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance '9cd89da4-d190-4bfc-81e1-da15b98d54ba' progress to 17 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 871.527755] env[62814]: DEBUG oslo_vmware.api [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Task: {'id': task-4293821, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230715} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.528035] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.528227] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 871.528403] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.528575] env[62814]: INFO nova.compute.manager [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Took 1.16 seconds to destroy the instance on the hypervisor. [ 871.528843] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 871.529045] env[62814]: DEBUG nova.compute.manager [-] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 871.529138] env[62814]: DEBUG nova.network.neutron [-] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 871.602919] env[62814]: INFO nova.compute.manager [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Rebuilding instance [ 871.630475] env[62814]: DEBUG nova.scheduler.client.report [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 871.654613] env[62814]: DEBUG nova.compute.manager [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 871.655738] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96029ef9-d5d6-40ea-8721-0886e550922a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.708875] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524436ec-587d-3f21-1c06-28e17114013e, 'name': SearchDatastore_Task, 'duration_secs': 0.009471} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.709726] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6384b8de-6fbc-4975-ac7a-685a8a677f7a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.715050] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 871.715050] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520b5e34-154f-bba4-250e-8fb07cafe3d6" [ 871.715050] env[62814]: _type = "Task" [ 871.715050] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.723028] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520b5e34-154f-bba4-250e-8fb07cafe3d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.805177] env[62814]: DEBUG nova.compute.manager [req-438ef202-9080-4774-ba66-7a38f2e65058 req-2b723dbc-e45f-4234-a454-5502586d584d service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Received event network-vif-deleted-5d635de7-0d82-4d28-a72c-35736e7d8f3b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 871.805426] env[62814]: INFO nova.compute.manager [req-438ef202-9080-4774-ba66-7a38f2e65058 req-2b723dbc-e45f-4234-a454-5502586d584d service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Neutron deleted interface 5d635de7-0d82-4d28-a72c-35736e7d8f3b; detaching it from the instance and deleting it from the info cache [ 871.807952] env[62814]: DEBUG nova.network.neutron [req-438ef202-9080-4774-ba66-7a38f2e65058 req-2b723dbc-e45f-4234-a454-5502586d584d service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Updating instance_info_cache with network_info: [{"id": "526f5c77-efba-4403-b12e-cc66f3cbd46d", "address": "fa:16:3e:8d:4e:bf", "network": {"id": "ca122834-ac3d-45cb-a8aa-34c46bfcf61a", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-214724823", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.101", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14adacd89bc460d8d1a7c7d2ecc5d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9f87a752-ebb0-49a4-a67b-e356fa45b89b", "external-id": "nsx-vlan-transportzone-889", "segmentation_id": 889, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap526f5c77-ef", "ovs_interfaceid": "526f5c77-efba-4403-b12e-cc66f3cbd46d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.950145] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 871.950522] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.950635] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 871.950819] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.950968] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 871.951129] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 871.951343] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 871.951502] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 871.951668] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 871.951834] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 871.952052] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 871.958336] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5008ed38-766e-44ff-871e-f68dd56ee0bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.974568] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 871.974568] env[62814]: value = "task-4293822" [ 871.974568] env[62814]: _type = "Task" [ 871.974568] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.985121] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293822, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.135655] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.669s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.136234] env[62814]: DEBUG nova.compute.manager [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 872.139639] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 34.013s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 872.139639] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 872.139828] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 872.140026] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.823s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 872.141733] env[62814]: INFO nova.compute.claims [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.146026] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef8cea0-ee02-4084-9d25-dc49809dfb73 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.153361] env[62814]: DEBUG nova.network.neutron [req-a32a7a9a-a83a-4b3e-978d-fa9aa9138269 req-c2fba204-e3ff-4d79-a226-ce56804983ef service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Updated VIF entry in instance network info cache for port 696f3530-9858-47e0-be99-e223f03b58fb. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.153705] env[62814]: DEBUG nova.network.neutron [req-a32a7a9a-a83a-4b3e-978d-fa9aa9138269 req-c2fba204-e3ff-4d79-a226-ce56804983ef service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Updating instance_info_cache with network_info: [{"id": "696f3530-9858-47e0-be99-e223f03b58fb", "address": "fa:16:3e:d4:fc:a6", "network": {"id": "7efb312b-bb3e-4806-a5cd-a2122e26982e", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-651938622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7f19ca23f48c40d391e7348ebf008751", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d986680e-ad16-45b1-bf6d-cd2fe661679f", "external-id": "nsx-vlan-transportzone-397", "segmentation_id": 397, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap696f3530-98", "ovs_interfaceid": "696f3530-9858-47e0-be99-e223f03b58fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.156197] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb3aec43-9b6f-4e00-b63f-2b36fbf03ce4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.176902] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b53edff-1c78-4319-9653-96d6bcd56cac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.185054] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590a2989-3c8b-4c82-b1dc-60d523309230 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.217903] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=177917MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 872.217903] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 872.225661] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520b5e34-154f-bba4-250e-8fb07cafe3d6, 'name': SearchDatastore_Task, 'duration_secs': 0.029637} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.225994] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.226303] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 9be62576-5a05-473f-befd-b33f5fde9185/9be62576-5a05-473f-befd-b33f5fde9185.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 872.226555] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a0bb754-595e-43d8-93f4-0732352174bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.233170] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 872.233170] env[62814]: value = "task-4293823" [ 872.233170] env[62814]: _type = "Task" [ 872.233170] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.240866] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293823, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.309452] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c30d77c1-5f40-4921-a38b-4ad2f268b7e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.319757] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109dbffe-aaf9-49c6-b6d2-7e865ac1c5c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.357830] env[62814]: DEBUG nova.compute.manager [req-438ef202-9080-4774-ba66-7a38f2e65058 req-2b723dbc-e45f-4234-a454-5502586d584d service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Detach interface failed, port_id=5d635de7-0d82-4d28-a72c-35736e7d8f3b, reason: Instance a473cc7e-0f86-4a18-9789-0aabfff430ab could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 872.485309] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293822, 'name': ReconfigVM_Task, 'duration_secs': 0.322114} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.486040] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance '9cd89da4-d190-4bfc-81e1-da15b98d54ba' progress to 33 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 872.522521] env[62814]: DEBUG nova.network.neutron [-] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.647187] env[62814]: DEBUG nova.compute.utils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 872.648724] env[62814]: DEBUG nova.compute.manager [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 872.648911] env[62814]: DEBUG nova.network.neutron [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 872.662378] env[62814]: DEBUG oslo_concurrency.lockutils [req-a32a7a9a-a83a-4b3e-978d-fa9aa9138269 req-c2fba204-e3ff-4d79-a226-ce56804983ef service nova] Releasing lock "refresh_cache-cd89e4d5-c08e-42c9-aea8-84c20714e196" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 872.681634] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.681978] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8791ecc8-c433-44be-8bbb-3e006b26077b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.691168] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 872.691168] env[62814]: value = "task-4293824" [ 872.691168] env[62814]: _type = "Task" [ 872.691168] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.699900] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293824, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.706739] env[62814]: DEBUG nova.policy [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '802c41e3db9f48b484ffc1ef687b79d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7240b2dcb8304ab68171223650d19172', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 872.748098] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293823, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.992989] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:36:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='82139a72-836d-436a-81f9-b33cf43b83c7',id=40,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1395970101',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 872.993413] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.993581] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 872.993770] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.993922] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 872.994077] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 872.994340] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 872.994542] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 872.994679] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 872.994980] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 872.995077] env[62814]: DEBUG nova.virt.hardware [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 873.000415] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Reconfiguring VM instance instance-00000030 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 873.000623] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddfb32b7-6955-452e-89a0-6320ee2d8f68 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.020675] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 873.020675] env[62814]: value = "task-4293825" [ 873.020675] env[62814]: _type = "Task" [ 873.020675] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.025981] env[62814]: INFO nova.compute.manager [-] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Took 1.50 seconds to deallocate network for instance. [ 873.035056] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293825, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.098150] env[62814]: DEBUG nova.network.neutron [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Successfully created port: 5099a0d1-867c-4372-a319-ad384543c523 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.154119] env[62814]: DEBUG nova.compute.manager [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 873.201311] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293824, 'name': PowerOffVM_Task, 'duration_secs': 0.299003} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.204061] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.204574] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.206166] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6fb3983-c857-43ff-8a7a-774fc577e0be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.213885] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.214149] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05e29ca6-08ee-416f-ae52-93283dd3754a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.243163] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293823, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.639014} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.243559] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 9be62576-5a05-473f-befd-b33f5fde9185/9be62576-5a05-473f-befd-b33f5fde9185.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 873.243791] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 873.244057] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-23f53342-2a2e-43e3-afb0-6421199652aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.249775] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 873.249775] env[62814]: value = "task-4293827" [ 873.249775] env[62814]: _type = "Task" [ 873.249775] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.257908] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293827, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.281126] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.281126] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.281126] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleting the datastore file [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.281126] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cd7bd2cb-d37d-413d-9690-e2129221cdff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.287950] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 873.287950] env[62814]: value = "task-4293828" [ 873.287950] env[62814]: _type = "Task" [ 873.287950] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.297448] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293828, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.529751] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293825, 'name': ReconfigVM_Task, 'duration_secs': 0.164256} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.532234] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Reconfigured VM instance instance-00000030 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 873.533201] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677c0bf9-276b-40ca-9b7d-ffec6e18bc68 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.536422] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 873.555866] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 9cd89da4-d190-4bfc-81e1-da15b98d54ba/9cd89da4-d190-4bfc-81e1-da15b98d54ba.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.558303] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98502d9f-cb77-4c29-afe6-c08f0fedcfbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.577092] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 873.577092] env[62814]: value = "task-4293829" [ 873.577092] env[62814]: _type = "Task" [ 873.577092] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.587102] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293829, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.647266] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2711cb-44eb-4f41-9f71-52acd8324c50 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.654292] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04eae88d-d7da-43c3-99b3-e10b24bb28e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.689533] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0d835c-1010-4591-8064-b8f5381ce021 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.697635] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51429656-3a03-44a9-acef-4cfc382130ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.711089] env[62814]: DEBUG nova.compute.provider_tree [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.759117] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293827, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071706} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.759340] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.760071] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b7a3f53-2048-4a78-be9a-badf2271e6c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.781776] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 9be62576-5a05-473f-befd-b33f5fde9185/9be62576-5a05-473f-befd-b33f5fde9185.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.782047] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5073241-5ac2-4547-b6fe-5e3742889be9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.804701] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293828, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.805917] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 873.805917] env[62814]: value = "task-4293830" [ 873.805917] env[62814]: _type = "Task" [ 873.805917] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.812985] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293830, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.832082] env[62814]: DEBUG nova.compute.manager [req-d927e1ae-da53-454d-aa9f-af8dea4a6662 req-a7bd1f34-243d-40ea-a980-8ced33e3c8eb service nova] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Received event network-vif-deleted-526f5c77-efba-4403-b12e-cc66f3cbd46d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 874.089284] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293829, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.167523] env[62814]: DEBUG nova.compute.manager [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 874.204427] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 874.204752] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.204925] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 874.205152] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.205318] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 874.205788] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 874.205788] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 874.205903] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 874.206049] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 874.206219] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 874.206483] env[62814]: DEBUG nova.virt.hardware [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 874.207396] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1547ba98-9ce7-4548-9edc-3d562ffa59b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.215237] env[62814]: DEBUG nova.scheduler.client.report [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 874.219784] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550641cc-0700-4e43-9484-ebd90ba8b29c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.306306] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293828, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.666319} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.306728] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.306928] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 874.307114] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.318118] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293830, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.526479] env[62814]: DEBUG nova.compute.manager [req-e0502168-48fe-4d68-b6e7-e8d2db9f2b50 req-3c968b88-7307-400c-88c4-34db2fa4bf35 service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Received event network-vif-plugged-5099a0d1-867c-4372-a319-ad384543c523 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 874.528697] env[62814]: DEBUG oslo_concurrency.lockutils [req-e0502168-48fe-4d68-b6e7-e8d2db9f2b50 req-3c968b88-7307-400c-88c4-34db2fa4bf35 service nova] Acquiring lock "844737d8-d852-44bb-bf9d-e673e737ef33-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.528697] env[62814]: DEBUG oslo_concurrency.lockutils [req-e0502168-48fe-4d68-b6e7-e8d2db9f2b50 req-3c968b88-7307-400c-88c4-34db2fa4bf35 service nova] Lock "844737d8-d852-44bb-bf9d-e673e737ef33-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.528697] env[62814]: DEBUG oslo_concurrency.lockutils [req-e0502168-48fe-4d68-b6e7-e8d2db9f2b50 req-3c968b88-7307-400c-88c4-34db2fa4bf35 service nova] Lock "844737d8-d852-44bb-bf9d-e673e737ef33-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.528697] env[62814]: DEBUG nova.compute.manager [req-e0502168-48fe-4d68-b6e7-e8d2db9f2b50 req-3c968b88-7307-400c-88c4-34db2fa4bf35 service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] No waiting events found dispatching network-vif-plugged-5099a0d1-867c-4372-a319-ad384543c523 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 874.528697] env[62814]: WARNING nova.compute.manager [req-e0502168-48fe-4d68-b6e7-e8d2db9f2b50 req-3c968b88-7307-400c-88c4-34db2fa4bf35 service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Received unexpected event network-vif-plugged-5099a0d1-867c-4372-a319-ad384543c523 for instance with vm_state building and task_state spawning. [ 874.587501] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293829, 'name': ReconfigVM_Task, 'duration_secs': 0.843167} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.588306] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 9cd89da4-d190-4bfc-81e1-da15b98d54ba/9cd89da4-d190-4bfc-81e1-da15b98d54ba.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.588765] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance '9cd89da4-d190-4bfc-81e1-da15b98d54ba' progress to 50 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 874.632570] env[62814]: DEBUG nova.network.neutron [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Successfully updated port: 5099a0d1-867c-4372-a319-ad384543c523 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.725921] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.586s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.726483] env[62814]: DEBUG nova.compute.manager [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 874.729196] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.866s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.729418] env[62814]: DEBUG nova.objects.instance [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lazy-loading 'resources' on Instance uuid 7cca1c35-6bfc-450d-ba74-0e825b160e8f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.818884] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293830, 'name': ReconfigVM_Task, 'duration_secs': 0.614325} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.819339] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 9be62576-5a05-473f-befd-b33f5fde9185/9be62576-5a05-473f-befd-b33f5fde9185.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.819647] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Creating Virtual Disk of size 1048576 KB and adapter type paraVirtual on the data store {{(pid=62814) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1383}} [ 874.820278] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CreateVirtualDisk_Task with opID=oslo.vmware-598a509e-e3b7-4505-b2fc-1ed57c9692d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.827162] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 874.827162] env[62814]: value = "task-4293831" [ 874.827162] env[62814]: _type = "Task" [ 874.827162] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.835836] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293831, 'name': CreateVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.101953] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86a2fec-6caf-48ac-96a5-4bf246cc3905 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.122745] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c27b21d5-cf5a-4da3-909f-b0af5d523e52 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.140654] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquiring lock "refresh_cache-844737d8-d852-44bb-bf9d-e673e737ef33" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.141314] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquired lock "refresh_cache-844737d8-d852-44bb-bf9d-e673e737ef33" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 875.141314] env[62814]: DEBUG nova.network.neutron [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 875.142197] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance '9cd89da4-d190-4bfc-81e1-da15b98d54ba' progress to 67 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 875.233347] env[62814]: DEBUG nova.compute.utils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 875.237700] env[62814]: DEBUG nova.compute.manager [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 875.238405] env[62814]: DEBUG nova.network.neutron [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 875.296930] env[62814]: DEBUG nova.policy [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd05b1b390e0489d91f4f386f4c6fe41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65a886f76f404ce5a6d0841241d8f120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 875.345966] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293831, 'name': CreateVirtualDisk_Task, 'duration_secs': 0.051446} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.346281] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Created Virtual Disk of size 1048576 KB and type thin {{(pid=62814) create_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1404}} [ 875.347173] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2314ac-7e57-49ba-9350-eb3334953a8b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.379283] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 9be62576-5a05-473f-befd-b33f5fde9185/ephemeral_0.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 875.383573] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 875.383819] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.383972] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 875.384223] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.384307] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 875.384462] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 875.384708] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 875.384877] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 875.385058] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 875.385224] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 875.385395] env[62814]: DEBUG nova.virt.hardware [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 875.386036] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-154d560a-c860-40d1-8ee0-f3379cdab087 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.399556] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54fe3be2-5fe5-4a43-918a-35544dd68b84 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.410186] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201b6387-7b7c-49e0-9de9-8c2d3230199c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.415026] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 875.415026] env[62814]: value = "task-4293832" [ 875.415026] env[62814]: _type = "Task" [ 875.415026] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.428910] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:cb:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ffcecdaa-a7b8-49fc-9371-dbdb7744688e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b7654ff-51f0-4a51-9b69-04d3352ddbee', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.437417] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 875.438801] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 875.438801] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6eadde8-f34b-48a8-be1b-adbf327888c1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.460925] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293832, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.466899] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 875.466899] env[62814]: value = "task-4293833" [ 875.466899] env[62814]: _type = "Task" [ 875.466899] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.474504] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293833, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.658171] env[62814]: DEBUG nova.network.neutron [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Successfully created port: d9875d9c-cdba-4f00-b6a3-72216a3f9f22 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.690515] env[62814]: DEBUG nova.network.neutron [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 875.734954] env[62814]: DEBUG nova.network.neutron [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Port bc38976c-8838-47e6-b4f8-cd013e62b15b binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 875.738443] env[62814]: DEBUG nova.compute.manager [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 875.816568] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa462f0-960f-46ff-927c-5980423975b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.825109] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739c2307-b630-4ccb-b797-c446c3492a05 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.859637] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e278a7-a548-44c7-91b2-0a080af83985 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.867246] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10045e4-1e5f-4241-9f89-4cac79cd657d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.883023] env[62814]: DEBUG nova.compute.provider_tree [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.893857] env[62814]: DEBUG nova.network.neutron [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Updating instance_info_cache with network_info: [{"id": "5099a0d1-867c-4372-a319-ad384543c523", "address": "fa:16:3e:e1:f6:e7", "network": {"id": "76bc0b08-444d-4ac0-9424-1d0cfa825dc7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-613849228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7240b2dcb8304ab68171223650d19172", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5099a0d1-86", "ovs_interfaceid": "5099a0d1-867c-4372-a319-ad384543c523", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.924580] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293832, 'name': ReconfigVM_Task, 'duration_secs': 0.310635} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.924860] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 9be62576-5a05-473f-befd-b33f5fde9185/ephemeral_0.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 875.925554] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a086c0a2-f8f7-49d0-8f2a-5de45fa1c23a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.932156] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 875.932156] env[62814]: value = "task-4293834" [ 875.932156] env[62814]: _type = "Task" [ 875.932156] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.941053] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293834, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.977655] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293833, 'name': CreateVM_Task, 'duration_secs': 0.311159} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.978227] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.978952] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.979157] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 875.979481] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 875.979744] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bec31633-ac17-40a0-8d8e-1a611cfb0147 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.984609] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 875.984609] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5211e2a1-33a9-adf5-9716-1024799c2b11" [ 875.984609] env[62814]: _type = "Task" [ 875.984609] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.994431] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5211e2a1-33a9-adf5-9716-1024799c2b11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.383904] env[62814]: DEBUG nova.scheduler.client.report [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 876.396609] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Releasing lock "refresh_cache-844737d8-d852-44bb-bf9d-e673e737ef33" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 876.396964] env[62814]: DEBUG nova.compute.manager [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Instance network_info: |[{"id": "5099a0d1-867c-4372-a319-ad384543c523", "address": "fa:16:3e:e1:f6:e7", "network": {"id": "76bc0b08-444d-4ac0-9424-1d0cfa825dc7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-613849228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7240b2dcb8304ab68171223650d19172", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5099a0d1-86", "ovs_interfaceid": "5099a0d1-867c-4372-a319-ad384543c523", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 876.397759] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:f6:e7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53915f38-d7a0-42ec-8b30-1eacfb2cc379', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5099a0d1-867c-4372-a319-ad384543c523', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.407168] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Creating folder: Project (7240b2dcb8304ab68171223650d19172). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 876.408090] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dea97f8c-851a-4919-a62a-03f73cb35820 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.419205] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Created folder: Project (7240b2dcb8304ab68171223650d19172) in parent group-v845547. [ 876.419324] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Creating folder: Instances. Parent ref: group-v845711. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 876.419519] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ae72c813-bdd8-4a42-8594-bc9ec8e9d287 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.429106] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Created folder: Instances in parent group-v845711. [ 876.429106] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 876.429356] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 876.429483] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e608b904-16c7-4ecf-8c7d-86d469b43483 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.454476] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.454476] env[62814]: value = "task-4293837" [ 876.454476] env[62814]: _type = "Task" [ 876.454476] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.457442] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293834, 'name': Rename_Task, 'duration_secs': 0.158448} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.460832] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 876.461092] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-091a603f-9c0d-4589-bc2b-e32a7bd9569b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.468843] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 876.468843] env[62814]: value = "task-4293838" [ 876.468843] env[62814]: _type = "Task" [ 876.468843] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.477148] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293838, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.493996] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5211e2a1-33a9-adf5-9716-1024799c2b11, 'name': SearchDatastore_Task, 'duration_secs': 0.011137} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.494335] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 876.494663] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.494846] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.494990] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 876.495185] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.495456] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f4affe8-53f8-49e7-95e6-e7b12d83330e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.504897] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.505237] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 876.506396] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c41659b2-80eb-4180-b6d7-f082cc397d34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.513150] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 876.513150] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209bdf4-6c63-6248-d0b3-d05be851f12b" [ 876.513150] env[62814]: _type = "Task" [ 876.513150] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.523364] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209bdf4-6c63-6248-d0b3-d05be851f12b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.554030] env[62814]: DEBUG nova.compute.manager [req-39e7f2ac-b53c-4945-a2c3-550fb6517764 req-6e502f8c-3da5-48b5-9038-3619ca1a434e service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Received event network-changed-5099a0d1-867c-4372-a319-ad384543c523 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 876.554387] env[62814]: DEBUG nova.compute.manager [req-39e7f2ac-b53c-4945-a2c3-550fb6517764 req-6e502f8c-3da5-48b5-9038-3619ca1a434e service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Refreshing instance network info cache due to event network-changed-5099a0d1-867c-4372-a319-ad384543c523. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 876.554749] env[62814]: DEBUG oslo_concurrency.lockutils [req-39e7f2ac-b53c-4945-a2c3-550fb6517764 req-6e502f8c-3da5-48b5-9038-3619ca1a434e service nova] Acquiring lock "refresh_cache-844737d8-d852-44bb-bf9d-e673e737ef33" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.555024] env[62814]: DEBUG oslo_concurrency.lockutils [req-39e7f2ac-b53c-4945-a2c3-550fb6517764 req-6e502f8c-3da5-48b5-9038-3619ca1a434e service nova] Acquired lock "refresh_cache-844737d8-d852-44bb-bf9d-e673e737ef33" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 876.555306] env[62814]: DEBUG nova.network.neutron [req-39e7f2ac-b53c-4945-a2c3-550fb6517764 req-6e502f8c-3da5-48b5-9038-3619ca1a434e service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Refreshing network info cache for port 5099a0d1-867c-4372-a319-ad384543c523 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 876.754527] env[62814]: DEBUG nova.compute.manager [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 876.769521] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 876.772749] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 876.772749] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.799566] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 876.799825] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.799982] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 876.800231] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.800424] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 876.800593] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 876.800844] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 876.801052] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 876.801253] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 876.801464] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 876.801686] env[62814]: DEBUG nova.virt.hardware [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 876.802910] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a6ae6d-c94f-48a5-811e-1e7c1137c55d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.812930] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c18a88-ca01-43ad-b09f-175b7c7d1c8e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.889858] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.161s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.892708] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.214s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 876.893216] env[62814]: DEBUG nova.objects.instance [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lazy-loading 'resources' on Instance uuid 413eaa2e-7bbc-402e-b0d3-f030b6395d7b {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.919790] env[62814]: INFO nova.scheduler.client.report [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Deleted allocations for instance 7cca1c35-6bfc-450d-ba74-0e825b160e8f [ 876.967044] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293837, 'name': CreateVM_Task, 'duration_secs': 0.31755} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.967195] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 876.967959] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.968152] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 876.968508] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 876.968737] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5113f90b-12e9-40ee-8d66-0a37dd84f1e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.976294] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for the task: (returnval){ [ 876.976294] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52420ebc-73dc-eafb-5a17-1bb025e5be81" [ 876.976294] env[62814]: _type = "Task" [ 876.976294] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.979958] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293838, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.989503] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52420ebc-73dc-eafb-5a17-1bb025e5be81, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.024878] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209bdf4-6c63-6248-d0b3-d05be851f12b, 'name': SearchDatastore_Task, 'duration_secs': 0.012956} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.025723] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80ddb97e-2442-4f5f-87da-8123528cece6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.030849] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 877.030849] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5221fa9f-a2a0-4027-8128-b663ec4225b5" [ 877.030849] env[62814]: _type = "Task" [ 877.030849] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.039536] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5221fa9f-a2a0-4027-8128-b663ec4225b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.273668] env[62814]: DEBUG nova.network.neutron [req-39e7f2ac-b53c-4945-a2c3-550fb6517764 req-6e502f8c-3da5-48b5-9038-3619ca1a434e service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Updated VIF entry in instance network info cache for port 5099a0d1-867c-4372-a319-ad384543c523. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 877.274018] env[62814]: DEBUG nova.network.neutron [req-39e7f2ac-b53c-4945-a2c3-550fb6517764 req-6e502f8c-3da5-48b5-9038-3619ca1a434e service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Updating instance_info_cache with network_info: [{"id": "5099a0d1-867c-4372-a319-ad384543c523", "address": "fa:16:3e:e1:f6:e7", "network": {"id": "76bc0b08-444d-4ac0-9424-1d0cfa825dc7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-613849228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7240b2dcb8304ab68171223650d19172", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5099a0d1-86", "ovs_interfaceid": "5099a0d1-867c-4372-a319-ad384543c523", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.319267] env[62814]: DEBUG nova.network.neutron [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Successfully updated port: d9875d9c-cdba-4f00-b6a3-72216a3f9f22 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.429986] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2a716-796c-48c5-acc0-6143124fd8ca tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "7cca1c35-6bfc-450d-ba74-0e825b160e8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.032s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 877.482398] env[62814]: DEBUG oslo_vmware.api [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4293838, 'name': PowerOnVM_Task, 'duration_secs': 0.529666} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.487428] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.487645] env[62814]: INFO nova.compute.manager [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Took 10.47 seconds to spawn the instance on the hypervisor. [ 877.487826] env[62814]: DEBUG nova.compute.manager [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 877.488730] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30599976-72c8-4895-824d-5b89e0e4cd93 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.496168] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52420ebc-73dc-eafb-5a17-1bb025e5be81, 'name': SearchDatastore_Task, 'duration_secs': 0.027594} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.499702] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.499934] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.500163] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.541724] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5221fa9f-a2a0-4027-8128-b663ec4225b5, 'name': SearchDatastore_Task, 'duration_secs': 0.00884} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.542641] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.543518] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.543518] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.543518] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.543518] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9971e824-f3e1-4c4b-ace1-5a1d2cd6d5e1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.547962] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4de18c1-9c8c-4d41-9b1e-dcd758dcb06e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.557412] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 877.557412] env[62814]: value = "task-4293839" [ 877.557412] env[62814]: _type = "Task" [ 877.557412] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.558795] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.558795] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 877.562220] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1d3fc9e-78df-4ce4-a5dc-39ccf7148c13 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.575410] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293839, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.577042] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for the task: (returnval){ [ 877.577042] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5237dd4c-bc47-de88-5ca7-1d85bf2b9317" [ 877.577042] env[62814]: _type = "Task" [ 877.577042] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.588593] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5237dd4c-bc47-de88-5ca7-1d85bf2b9317, 'name': SearchDatastore_Task, 'duration_secs': 0.009071} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.589382] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb526bdf-99ab-4585-9dff-2ec55322f482 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.595951] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for the task: (returnval){ [ 877.595951] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f0ec34-9ce5-3946-719d-538caa5fae99" [ 877.595951] env[62814]: _type = "Task" [ 877.595951] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.603793] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f0ec34-9ce5-3946-719d-538caa5fae99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.781840] env[62814]: DEBUG oslo_concurrency.lockutils [req-39e7f2ac-b53c-4945-a2c3-550fb6517764 req-6e502f8c-3da5-48b5-9038-3619ca1a434e service nova] Releasing lock "refresh_cache-844737d8-d852-44bb-bf9d-e673e737ef33" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 877.814281] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.814600] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.814862] env[62814]: DEBUG nova.network.neutron [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.821389] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "refresh_cache-1e259ec6-d31a-453b-87e0-baa446665d56" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.821795] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "refresh_cache-1e259ec6-d31a-453b-87e0-baa446665d56" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 877.821979] env[62814]: DEBUG nova.network.neutron [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.857558] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6c6239-4fd2-489b-b37b-54e6d88c3426 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.865623] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8808ae93-88fe-4b21-928e-633859715cb2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.899554] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5261720d-4fc1-421a-afdd-f6aef427e0b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.909881] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6a4860-33a1-4509-ab52-5fed94864626 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.926439] env[62814]: DEBUG nova.compute.provider_tree [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.011425] env[62814]: INFO nova.compute.manager [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Took 53.83 seconds to build instance. [ 878.067761] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293839, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49185} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.068035] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 878.068247] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.068520] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6cc3efe8-7f6d-4a19-8587-046f0e08ee4e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.074761] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 878.074761] env[62814]: value = "task-4293840" [ 878.074761] env[62814]: _type = "Task" [ 878.074761] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.082975] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293840, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.105609] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f0ec34-9ce5-3946-719d-538caa5fae99, 'name': SearchDatastore_Task, 'duration_secs': 0.00866} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.105875] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.106153] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 844737d8-d852-44bb-bf9d-e673e737ef33/844737d8-d852-44bb-bf9d-e673e737ef33.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 878.106406] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ca8bab8-c5f2-49a1-8814-bbec87be4365 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.112462] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for the task: (returnval){ [ 878.112462] env[62814]: value = "task-4293841" [ 878.112462] env[62814]: _type = "Task" [ 878.112462] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.119824] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293841, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.354859] env[62814]: DEBUG nova.network.neutron [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.430161] env[62814]: DEBUG nova.scheduler.client.report [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 878.513633] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4676491b-f7ea-46d7-8ee3-2e25a52bdd12 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "9be62576-5a05-473f-befd-b33f5fde9185" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 55.992s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.568702] env[62814]: DEBUG nova.network.neutron [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Updating instance_info_cache with network_info: [{"id": "d9875d9c-cdba-4f00-b6a3-72216a3f9f22", "address": "fa:16:3e:6b:c2:b2", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9875d9c-cd", "ovs_interfaceid": "d9875d9c-cdba-4f00-b6a3-72216a3f9f22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.592468] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293840, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089937} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.594018] env[62814]: DEBUG nova.network.neutron [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance_info_cache with network_info: [{"id": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "address": "fa:16:3e:fe:5b:6c", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38976c-88", "ovs_interfaceid": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.595172] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.596441] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-785fb1ae-ea2d-45b3-98c0-c79838c1a64f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.625762] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Reconfiguring VM instance instance-00000016 to attach disk [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.632023] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fa4b88a-ce10-40c4-b338-57cdef863ded {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.652344] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293841, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.654111] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 878.654111] env[62814]: value = "task-4293842" [ 878.654111] env[62814]: _type = "Task" [ 878.654111] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.665031] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293842, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.715343] env[62814]: DEBUG nova.compute.manager [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Received event network-vif-plugged-d9875d9c-cdba-4f00-b6a3-72216a3f9f22 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 878.715626] env[62814]: DEBUG oslo_concurrency.lockutils [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] Acquiring lock "1e259ec6-d31a-453b-87e0-baa446665d56-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.715863] env[62814]: DEBUG oslo_concurrency.lockutils [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] Lock "1e259ec6-d31a-453b-87e0-baa446665d56-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.716048] env[62814]: DEBUG oslo_concurrency.lockutils [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] Lock "1e259ec6-d31a-453b-87e0-baa446665d56-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.716223] env[62814]: DEBUG nova.compute.manager [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] No waiting events found dispatching network-vif-plugged-d9875d9c-cdba-4f00-b6a3-72216a3f9f22 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 878.716404] env[62814]: WARNING nova.compute.manager [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Received unexpected event network-vif-plugged-d9875d9c-cdba-4f00-b6a3-72216a3f9f22 for instance with vm_state building and task_state spawning. [ 878.717057] env[62814]: DEBUG nova.compute.manager [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Received event network-changed-d9875d9c-cdba-4f00-b6a3-72216a3f9f22 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 878.717057] env[62814]: DEBUG nova.compute.manager [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Refreshing instance network info cache due to event network-changed-d9875d9c-cdba-4f00-b6a3-72216a3f9f22. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 878.717057] env[62814]: DEBUG oslo_concurrency.lockutils [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] Acquiring lock "refresh_cache-1e259ec6-d31a-453b-87e0-baa446665d56" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.938553] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.045s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.940736] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 33.514s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.942241] env[62814]: INFO nova.compute.claims [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.971837] env[62814]: INFO nova.scheduler.client.report [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Deleted allocations for instance 413eaa2e-7bbc-402e-b0d3-f030b6395d7b [ 879.030644] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "99864c68-5f11-4119-b8e3-3aa8719f267b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 879.030864] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "99864c68-5f11-4119-b8e3-3aa8719f267b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 879.078298] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "refresh_cache-1e259ec6-d31a-453b-87e0-baa446665d56" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 879.078648] env[62814]: DEBUG nova.compute.manager [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Instance network_info: |[{"id": "d9875d9c-cdba-4f00-b6a3-72216a3f9f22", "address": "fa:16:3e:6b:c2:b2", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9875d9c-cd", "ovs_interfaceid": "d9875d9c-cdba-4f00-b6a3-72216a3f9f22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 879.078949] env[62814]: DEBUG oslo_concurrency.lockutils [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] Acquired lock "refresh_cache-1e259ec6-d31a-453b-87e0-baa446665d56" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 879.079138] env[62814]: DEBUG nova.network.neutron [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Refreshing network info cache for port d9875d9c-cdba-4f00-b6a3-72216a3f9f22 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.080316] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:c2:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7e0240aa-a694-48fc-a0f9-6f2d3e71aa12', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd9875d9c-cdba-4f00-b6a3-72216a3f9f22', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 879.087735] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Creating folder: Project (65a886f76f404ce5a6d0841241d8f120). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 879.088636] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d587282-df42-4138-ad04-0adf0f018079 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.096632] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 879.102147] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Created folder: Project (65a886f76f404ce5a6d0841241d8f120) in parent group-v845547. [ 879.102332] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Creating folder: Instances. Parent ref: group-v845714. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 879.102759] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2a4ea329-6166-4567-a76a-b552026c6dfd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.111206] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Created folder: Instances in parent group-v845714. [ 879.111443] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 879.111608] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 879.111798] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-16a7b22c-b51d-4d9d-bd72-2aa86d3e89b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.135953] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293841, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.683325} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.137138] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 844737d8-d852-44bb-bf9d-e673e737ef33/844737d8-d852-44bb-bf9d-e673e737ef33.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 879.137356] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 879.137584] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.137584] env[62814]: value = "task-4293845" [ 879.137584] env[62814]: _type = "Task" [ 879.137584] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.137756] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41ae6ce8-0837-4b76-9ba6-2089b2f21b72 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.148309] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293845, 'name': CreateVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.152134] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for the task: (returnval){ [ 879.152134] env[62814]: value = "task-4293846" [ 879.152134] env[62814]: _type = "Task" [ 879.152134] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.161344] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293846, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.166987] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293842, 'name': ReconfigVM_Task, 'duration_secs': 0.349247} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.170481] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Reconfigured VM instance instance-00000016 to attach disk [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574/4ed66b36-b6c6-4673-9c03-169a01134574.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.171216] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf6ec5de-2526-492d-825c-94052f89c12a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.176762] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 879.176762] env[62814]: value = "task-4293847" [ 879.176762] env[62814]: _type = "Task" [ 879.176762] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.186126] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293847, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.411212] env[62814]: DEBUG nova.compute.manager [req-dfb98577-9c38-448e-a1b2-6fb152c6eddf req-30ed13e6-d2fb-4c1d-9dcb-09778faa660a service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Received event network-changed-5ec47315-906b-4ad9-a8c8-87622cb99c27 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 879.411212] env[62814]: DEBUG nova.compute.manager [req-dfb98577-9c38-448e-a1b2-6fb152c6eddf req-30ed13e6-d2fb-4c1d-9dcb-09778faa660a service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Refreshing instance network info cache due to event network-changed-5ec47315-906b-4ad9-a8c8-87622cb99c27. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 879.411212] env[62814]: DEBUG oslo_concurrency.lockutils [req-dfb98577-9c38-448e-a1b2-6fb152c6eddf req-30ed13e6-d2fb-4c1d-9dcb-09778faa660a service nova] Acquiring lock "refresh_cache-9be62576-5a05-473f-befd-b33f5fde9185" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.411645] env[62814]: DEBUG oslo_concurrency.lockutils [req-dfb98577-9c38-448e-a1b2-6fb152c6eddf req-30ed13e6-d2fb-4c1d-9dcb-09778faa660a service nova] Acquired lock "refresh_cache-9be62576-5a05-473f-befd-b33f5fde9185" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 879.411645] env[62814]: DEBUG nova.network.neutron [req-dfb98577-9c38-448e-a1b2-6fb152c6eddf req-30ed13e6-d2fb-4c1d-9dcb-09778faa660a service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Refreshing network info cache for port 5ec47315-906b-4ad9-a8c8-87622cb99c27 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.480831] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6277c0cf-b017-4dba-ba0a-aa977912dd06 tempest-ServerMetadataNegativeTestJSON-670324559 tempest-ServerMetadataNegativeTestJSON-670324559-project-member] Lock "413eaa2e-7bbc-402e-b0d3-f030b6395d7b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.252s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.535736] env[62814]: DEBUG nova.compute.manager [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 879.624464] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a584d168-7aa9-44c9-a52c-3f81d3195c9b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.654603] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b634f5d8-3ae5-4ba8-904f-4be38e3ca7a2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.667330] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance '9cd89da4-d190-4bfc-81e1-da15b98d54ba' progress to 83 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 879.678029] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293845, 'name': CreateVM_Task, 'duration_secs': 0.354181} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.682028] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293846, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10619} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.682028] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 879.682630] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.683393] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.683576] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 879.683894] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 879.684677] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11be27a5-5aad-4741-8bb2-98248baaf6a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.687252] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b98753e7-ff12-469e-a73f-0596465037a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.694123] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293847, 'name': Rename_Task, 'duration_secs': 0.135678} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.708079] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 879.718084] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 844737d8-d852-44bb-bf9d-e673e737ef33/844737d8-d852-44bb-bf9d-e673e737ef33.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.719105] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 879.719105] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c4c9b3-b8e7-e8db-92b5-0ca668fd7f25" [ 879.719105] env[62814]: _type = "Task" [ 879.719105] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.719105] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e21ec72c-09ce-42bf-9dba-b5a30c081fa7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.720531] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6028c3b1-a2e4-49b4-9ac4-e8bbde1bb436 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.749052] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c4c9b3-b8e7-e8db-92b5-0ca668fd7f25, 'name': SearchDatastore_Task, 'duration_secs': 0.013428} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.751013] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 879.751289] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.751531] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.751682] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 879.751869] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.752239] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for the task: (returnval){ [ 879.752239] env[62814]: value = "task-4293849" [ 879.752239] env[62814]: _type = "Task" [ 879.752239] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.752480] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 879.752480] env[62814]: value = "task-4293848" [ 879.752480] env[62814]: _type = "Task" [ 879.752480] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.752705] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed3ad3ad-8381-4b28-8f5f-b68fb5d02219 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.775427] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293849, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.775707] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293848, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.778044] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.778044] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 879.778159] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29091944-5d8d-4b2f-be79-deb5c6028b99 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.784726] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 879.784726] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5240780d-2c76-4bfc-411b-33c4b80d28b4" [ 879.784726] env[62814]: _type = "Task" [ 879.784726] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.795798] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5240780d-2c76-4bfc-411b-33c4b80d28b4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.938137] env[62814]: DEBUG nova.network.neutron [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Updated VIF entry in instance network info cache for port d9875d9c-cdba-4f00-b6a3-72216a3f9f22. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 879.938643] env[62814]: DEBUG nova.network.neutron [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Updating instance_info_cache with network_info: [{"id": "d9875d9c-cdba-4f00-b6a3-72216a3f9f22", "address": "fa:16:3e:6b:c2:b2", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9875d9c-cd", "ovs_interfaceid": "d9875d9c-cdba-4f00-b6a3-72216a3f9f22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.063772] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.181817] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.182195] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8b13529-533a-4360-81e1-96f38b1eedab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.190324] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 880.190324] env[62814]: value = "task-4293850" [ 880.190324] env[62814]: _type = "Task" [ 880.190324] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.204961] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293850, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.269353] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293848, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.273852] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293849, 'name': ReconfigVM_Task, 'duration_secs': 0.453042} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.274400] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 844737d8-d852-44bb-bf9d-e673e737ef33/844737d8-d852-44bb-bf9d-e673e737ef33.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.277272] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57cca38e-0a02-4e85-a9ce-c57539eb192b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.283954] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for the task: (returnval){ [ 880.283954] env[62814]: value = "task-4293851" [ 880.283954] env[62814]: _type = "Task" [ 880.283954] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.299387] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293851, 'name': Rename_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.302916] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5240780d-2c76-4bfc-411b-33c4b80d28b4, 'name': SearchDatastore_Task, 'duration_secs': 0.010396} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.306157] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bd3fcfe-46c6-437b-84ef-c92afecd859c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.311381] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 880.311381] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5219b29f-99b0-6c88-522e-8d2d38bd7bf5" [ 880.311381] env[62814]: _type = "Task" [ 880.311381] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.319326] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5219b29f-99b0-6c88-522e-8d2d38bd7bf5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.367699] env[62814]: DEBUG nova.network.neutron [req-dfb98577-9c38-448e-a1b2-6fb152c6eddf req-30ed13e6-d2fb-4c1d-9dcb-09778faa660a service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Updated VIF entry in instance network info cache for port 5ec47315-906b-4ad9-a8c8-87622cb99c27. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.368151] env[62814]: DEBUG nova.network.neutron [req-dfb98577-9c38-448e-a1b2-6fb152c6eddf req-30ed13e6-d2fb-4c1d-9dcb-09778faa660a service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Updating instance_info_cache with network_info: [{"id": "5ec47315-906b-4ad9-a8c8-87622cb99c27", "address": "fa:16:3e:5f:99:cd", "network": {"id": "b4436b99-a4a3-46c9-918a-4e6f6f6bb55e", "bridge": "br-int", "label": "tempest-ServersWithSpecificFlavorTestJSON-2020231604-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9d51711cd907404b8d0a48d904b599c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98d96b75-ac36-499a-adc2-130c8c1d55ca", "external-id": "nsx-vlan-transportzone-564", "segmentation_id": 564, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ec47315-90", "ovs_interfaceid": "5ec47315-906b-4ad9-a8c8-87622cb99c27", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.441414] env[62814]: DEBUG oslo_concurrency.lockutils [req-4ffd79c4-1618-4620-8569-2f08f68cf150 req-3ab2640d-a422-4bef-8bc5-aae15253eb37 service nova] Releasing lock "refresh_cache-1e259ec6-d31a-453b-87e0-baa446665d56" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 880.450199] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08f7e97c-e826-4c67-ae0d-e535ef8a79cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.458102] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf6998d-23ce-4415-889f-898348a08e4b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.490814] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be57f64-2118-442d-8894-6fb027f8b245 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.498897] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e284ec-6450-4110-819a-43c202ab7ca3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.513954] env[62814]: DEBUG nova.compute.provider_tree [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.702376] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293850, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.766237] env[62814]: DEBUG oslo_vmware.api [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293848, 'name': PowerOnVM_Task, 'duration_secs': 0.661054} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.766516] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 880.766721] env[62814]: DEBUG nova.compute.manager [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 880.767501] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7fe72b-a084-46bb-8596-b422127b91e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.794259] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293851, 'name': Rename_Task, 'duration_secs': 0.213891} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.794550] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.794794] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-77920319-7cf4-4a4c-9ff3-e2aaafaf5c4e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.802241] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for the task: (returnval){ [ 880.802241] env[62814]: value = "task-4293852" [ 880.802241] env[62814]: _type = "Task" [ 880.802241] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.811042] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293852, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.820727] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5219b29f-99b0-6c88-522e-8d2d38bd7bf5, 'name': SearchDatastore_Task, 'duration_secs': 0.009726} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.822722] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 880.822722] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1e259ec6-d31a-453b-87e0-baa446665d56/1e259ec6-d31a-453b-87e0-baa446665d56.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.822722] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbef7600-a95c-4679-917c-12304a311ed2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.830027] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 880.830027] env[62814]: value = "task-4293853" [ 880.830027] env[62814]: _type = "Task" [ 880.830027] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.838461] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.873652] env[62814]: DEBUG oslo_concurrency.lockutils [req-dfb98577-9c38-448e-a1b2-6fb152c6eddf req-30ed13e6-d2fb-4c1d-9dcb-09778faa660a service nova] Releasing lock "refresh_cache-9be62576-5a05-473f-befd-b33f5fde9185" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 881.018288] env[62814]: DEBUG nova.scheduler.client.report [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 881.202446] env[62814]: DEBUG oslo_vmware.api [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293850, 'name': PowerOnVM_Task, 'duration_secs': 0.922102} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.202810] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 881.203011] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f0b958fa-4841-4961-b0eb-36de34794884 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance '9cd89da4-d190-4bfc-81e1-da15b98d54ba' progress to 100 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 881.287070] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 881.314577] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293852, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.340783] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293853, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.525646] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.585s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 881.526328] env[62814]: DEBUG nova.compute.manager [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 881.529267] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.486s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 881.532839] env[62814]: INFO nova.compute.claims [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.813015] env[62814]: DEBUG oslo_vmware.api [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4293852, 'name': PowerOnVM_Task, 'duration_secs': 0.825032} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.813315] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 881.813516] env[62814]: INFO nova.compute.manager [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Took 7.65 seconds to spawn the instance on the hypervisor. [ 881.813696] env[62814]: DEBUG nova.compute.manager [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 881.814495] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5799a952-cc4e-4057-bba2-84ee216cdb87 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.842253] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293853, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.744296} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.842525] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1e259ec6-d31a-453b-87e0-baa446665d56/1e259ec6-d31a-453b-87e0-baa446665d56.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.842763] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.843037] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dd542b10-1866-4732-987f-647049d7765a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.851199] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 881.851199] env[62814]: value = "task-4293854" [ 881.851199] env[62814]: _type = "Task" [ 881.851199] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.864469] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293854, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.036952] env[62814]: DEBUG nova.compute.utils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 882.040606] env[62814]: DEBUG nova.compute.manager [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Not allocating networking since 'none' was specified. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 882.333042] env[62814]: INFO nova.compute.manager [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Took 51.24 seconds to build instance. [ 882.362162] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293854, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118639} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.362540] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.363348] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c425d401-f152-42d9-9823-20a875198843 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.386591] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 1e259ec6-d31a-453b-87e0-baa446665d56/1e259ec6-d31a-453b-87e0-baa446665d56.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.387252] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f19728d3-4a59-4995-bc2d-bb2bdce169cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.408531] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 882.408531] env[62814]: value = "task-4293855" [ 882.408531] env[62814]: _type = "Task" [ 882.408531] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.417584] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.541308] env[62814]: DEBUG nova.compute.manager [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 882.779089] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.779458] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.779806] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.780138] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.780544] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 882.784335] env[62814]: INFO nova.compute.manager [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Terminating instance [ 882.835837] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c366e5aa-1bc1-4c01-8c09-df09dfe6fddf tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lock "844737d8-d852-44bb-bf9d-e673e737ef33" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.880s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 882.917943] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293855, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.073164] env[62814]: DEBUG nova.compute.manager [req-4579d8be-3000-4345-b0f1-275c92c85c43 req-37090b17-5e45-4478-98c7-2b94a62ce7ff service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Received event network-changed-5099a0d1-867c-4372-a319-ad384543c523 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 883.073164] env[62814]: DEBUG nova.compute.manager [req-4579d8be-3000-4345-b0f1-275c92c85c43 req-37090b17-5e45-4478-98c7-2b94a62ce7ff service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Refreshing instance network info cache due to event network-changed-5099a0d1-867c-4372-a319-ad384543c523. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 883.073164] env[62814]: DEBUG oslo_concurrency.lockutils [req-4579d8be-3000-4345-b0f1-275c92c85c43 req-37090b17-5e45-4478-98c7-2b94a62ce7ff service nova] Acquiring lock "refresh_cache-844737d8-d852-44bb-bf9d-e673e737ef33" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.073164] env[62814]: DEBUG oslo_concurrency.lockutils [req-4579d8be-3000-4345-b0f1-275c92c85c43 req-37090b17-5e45-4478-98c7-2b94a62ce7ff service nova] Acquired lock "refresh_cache-844737d8-d852-44bb-bf9d-e673e737ef33" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 883.073421] env[62814]: DEBUG nova.network.neutron [req-4579d8be-3000-4345-b0f1-275c92c85c43 req-37090b17-5e45-4478-98c7-2b94a62ce7ff service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Refreshing network info cache for port 5099a0d1-867c-4372-a319-ad384543c523 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.135309] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54e752c-1a5d-4ec5-a9d7-f9ea52b01cb2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.146499] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ced5147-317c-4ce1-ab02-902d39ddccd0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.178309] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b3570d-07fc-4c6b-b75d-d78b5c81ae1e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.186156] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca067b70-9f0c-49a3-b027-0a113117bc46 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.200562] env[62814]: DEBUG nova.compute.provider_tree [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.289381] env[62814]: DEBUG nova.compute.manager [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 883.292107] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 883.292107] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6ff2ce-32d9-4473-ac24-60778cc6a387 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.304671] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.304671] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b6e32c32-c5e0-45f1-b325-b0c45de40fd6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.310513] env[62814]: DEBUG oslo_vmware.api [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 883.310513] env[62814]: value = "task-4293856" [ 883.310513] env[62814]: _type = "Task" [ 883.310513] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.331557] env[62814]: DEBUG oslo_vmware.api [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293856, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.418078] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293855, 'name': ReconfigVM_Task, 'duration_secs': 0.635343} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.418492] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 1e259ec6-d31a-453b-87e0-baa446665d56/1e259ec6-d31a-453b-87e0-baa446665d56.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.419164] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1b979d70-5a82-4aac-8f5c-4f3e6842ecf6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.426296] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 883.426296] env[62814]: value = "task-4293857" [ 883.426296] env[62814]: _type = "Task" [ 883.426296] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.434810] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293857, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.555587] env[62814]: DEBUG nova.compute.manager [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 883.582632] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 883.583425] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.583425] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 883.583425] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.583425] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 883.583603] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 883.583821] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 883.583985] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 883.584167] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 883.584329] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 883.584600] env[62814]: DEBUG nova.virt.hardware [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 883.585901] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2553fe-63ce-4a52-b59a-24bf92fd6f58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.597910] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca5eb39-0b6c-4a80-a599-d3d737363c1c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.615753] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 883.620567] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Creating folder: Project (83ad323be5b64f34afc26e9d0369cb9c). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 883.623520] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-490a9b5f-3d4d-46de-9a65-4f5465609150 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.633471] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Created folder: Project (83ad323be5b64f34afc26e9d0369cb9c) in parent group-v845547. [ 883.637016] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Creating folder: Instances. Parent ref: group-v845717. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 883.637016] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90321f07-7636-45fb-8604-919fa5d6a0aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.643181] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Created folder: Instances in parent group-v845717. [ 883.643425] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 883.643617] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 883.643824] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e74e079f-a0fb-4c66-961e-6e2a2dd58822 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.662862] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 883.662862] env[62814]: value = "task-4293860" [ 883.662862] env[62814]: _type = "Task" [ 883.662862] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.670569] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293860, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.704378] env[62814]: DEBUG nova.scheduler.client.report [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 883.821954] env[62814]: DEBUG oslo_vmware.api [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293856, 'name': PowerOffVM_Task, 'duration_secs': 0.194052} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.822101] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 883.822298] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 883.822568] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fdcff42e-a93d-4178-9078-243bc5b149d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.885755] env[62814]: DEBUG nova.network.neutron [req-4579d8be-3000-4345-b0f1-275c92c85c43 req-37090b17-5e45-4478-98c7-2b94a62ce7ff service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Updated VIF entry in instance network info cache for port 5099a0d1-867c-4372-a319-ad384543c523. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 883.886168] env[62814]: DEBUG nova.network.neutron [req-4579d8be-3000-4345-b0f1-275c92c85c43 req-37090b17-5e45-4478-98c7-2b94a62ce7ff service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Updating instance_info_cache with network_info: [{"id": "5099a0d1-867c-4372-a319-ad384543c523", "address": "fa:16:3e:e1:f6:e7", "network": {"id": "76bc0b08-444d-4ac0-9424-1d0cfa825dc7", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-613849228-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7240b2dcb8304ab68171223650d19172", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53915f38-d7a0-42ec-8b30-1eacfb2cc379", "external-id": "nsx-vlan-transportzone-928", "segmentation_id": 928, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5099a0d1-86", "ovs_interfaceid": "5099a0d1-867c-4372-a319-ad384543c523", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.901062] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 883.901062] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 883.901062] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleting the datastore file [datastore2] ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.901289] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7172f9c9-25b8-4923-ac58-595a875eaec7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.912428] env[62814]: DEBUG oslo_vmware.api [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 883.912428] env[62814]: value = "task-4293862" [ 883.912428] env[62814]: _type = "Task" [ 883.912428] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.922363] env[62814]: DEBUG oslo_vmware.api [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293862, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.938356] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293857, 'name': Rename_Task, 'duration_secs': 0.148711} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.938356] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 883.938356] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dde2cbd-f76f-46aa-b073-90423f7b466d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.946796] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 883.946796] env[62814]: value = "task-4293863" [ 883.946796] env[62814]: _type = "Task" [ 883.946796] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.954975] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.963448] env[62814]: DEBUG nova.network.neutron [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Port bc38976c-8838-47e6-b4f8-cd013e62b15b binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 883.963448] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.963448] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 883.963626] env[62814]: DEBUG nova.network.neutron [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.172245] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293860, 'name': CreateVM_Task, 'duration_secs': 0.325706} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.172482] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 884.172964] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.173244] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 884.173980] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 884.173980] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0031311-03f4-4650-9ce6-a21f1ba785e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.179078] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 884.179078] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528310f9-aa5d-80cc-531e-012a8bda9172" [ 884.179078] env[62814]: _type = "Task" [ 884.179078] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.187126] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528310f9-aa5d-80cc-531e-012a8bda9172, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.210186] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.681s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 884.210768] env[62814]: DEBUG nova.compute.manager [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 884.213473] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.693s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 884.214898] env[62814]: INFO nova.compute.claims [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.221355] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 884.221591] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 884.389317] env[62814]: DEBUG oslo_concurrency.lockutils [req-4579d8be-3000-4345-b0f1-275c92c85c43 req-37090b17-5e45-4478-98c7-2b94a62ce7ff service nova] Releasing lock "refresh_cache-844737d8-d852-44bb-bf9d-e673e737ef33" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 884.423109] env[62814]: DEBUG oslo_vmware.api [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293862, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.213204} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.423109] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.423291] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 884.423491] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 884.423718] env[62814]: INFO nova.compute.manager [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Took 1.13 seconds to destroy the instance on the hypervisor. [ 884.424011] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 884.424263] env[62814]: DEBUG nova.compute.manager [-] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 884.424387] env[62814]: DEBUG nova.network.neutron [-] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 884.457222] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293863, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.693107] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528310f9-aa5d-80cc-531e-012a8bda9172, 'name': SearchDatastore_Task, 'duration_secs': 0.009987} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.693499] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 884.693675] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 884.693936] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.694125] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 884.694268] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.695876] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1c091fe-dc7d-4567-b2fa-36b68771e5df {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.707040] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.707040] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 884.707040] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd6b2e1c-6fe7-4296-8b21-8520a327814b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.712530] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 884.712530] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529e7acc-88d3-4304-f1b3-dd2c23df04d5" [ 884.712530] env[62814]: _type = "Task" [ 884.712530] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.723945] env[62814]: DEBUG nova.compute.utils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 884.727973] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529e7acc-88d3-4304-f1b3-dd2c23df04d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.728587] env[62814]: DEBUG nova.compute.manager [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 884.734193] env[62814]: DEBUG nova.compute.manager [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 884.734352] env[62814]: DEBUG nova.network.neutron [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 884.797696] env[62814]: DEBUG nova.network.neutron [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance_info_cache with network_info: [{"id": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "address": "fa:16:3e:fe:5b:6c", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38976c-88", "ovs_interfaceid": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.831850] env[62814]: DEBUG nova.policy [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '67e95624a1234848a158decc960c5ef8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ef503fd02adc475fa79dda8b379b6514', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 884.962574] env[62814]: DEBUG oslo_vmware.api [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4293863, 'name': PowerOnVM_Task, 'duration_secs': 0.572179} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.962574] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 884.962867] env[62814]: INFO nova.compute.manager [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Took 8.21 seconds to spawn the instance on the hypervisor. [ 884.963102] env[62814]: DEBUG nova.compute.manager [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 884.963925] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffea225-0115-4d62-8cbe-29985561ca63 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.103175] env[62814]: DEBUG nova.compute.manager [req-de874647-bfa0-457f-adc3-bcb0bd8c1d75 req-9149f07c-449b-4e1b-9fba-b1d1b29ca685 service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Received event network-vif-deleted-aa760165-654c-49b6-97bb-5fc792c2071d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 885.103389] env[62814]: INFO nova.compute.manager [req-de874647-bfa0-457f-adc3-bcb0bd8c1d75 req-9149f07c-449b-4e1b-9fba-b1d1b29ca685 service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Neutron deleted interface aa760165-654c-49b6-97bb-5fc792c2071d; detaching it from the instance and deleting it from the info cache [ 885.103607] env[62814]: DEBUG nova.network.neutron [req-de874647-bfa0-457f-adc3-bcb0bd8c1d75 req-9149f07c-449b-4e1b-9fba-b1d1b29ca685 service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.225191] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529e7acc-88d3-4304-f1b3-dd2c23df04d5, 'name': SearchDatastore_Task, 'duration_secs': 0.018314} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.225893] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-021cab59-4ca3-4821-b494-657944a8b30d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.233689] env[62814]: DEBUG nova.network.neutron [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Successfully created port: 5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.237758] env[62814]: DEBUG nova.compute.manager [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 885.244900] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 885.244900] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522c8080-d43a-7e47-3cb1-01b22b652524" [ 885.244900] env[62814]: _type = "Task" [ 885.244900] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.256569] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522c8080-d43a-7e47-3cb1-01b22b652524, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.260500] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.301728] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 885.314692] env[62814]: DEBUG nova.network.neutron [-] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.484024] env[62814]: INFO nova.compute.manager [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Took 45.19 seconds to build instance. [ 885.608115] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d3802e82-0269-4e09-8eda-4983174068b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.619245] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ffc8e8-ef2e-4af0-9161-117a08f5bb0e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.662893] env[62814]: DEBUG nova.compute.manager [req-de874647-bfa0-457f-adc3-bcb0bd8c1d75 req-9149f07c-449b-4e1b-9fba-b1d1b29ca685 service nova] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Detach interface failed, port_id=aa760165-654c-49b6-97bb-5fc792c2071d, reason: Instance ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 885.734140] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f374043b-81b3-4f5e-9692-5687ad201074 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.742868] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1324897b-763e-491d-8a31-bce39df3f749 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.786564] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12d0999f-1040-4172-8b09-2b731cc520b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.794038] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522c8080-d43a-7e47-3cb1-01b22b652524, 'name': SearchDatastore_Task, 'duration_secs': 0.017034} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.794932] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 885.795252] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 885.796019] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a2226da-5ae5-4fcc-ba9f-ce599e90ff53 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.802435] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1e0e6d-1863-42b9-8901-f1653086aba1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.809702] env[62814]: DEBUG nova.compute.manager [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62814) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:901}} [ 885.809702] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.809702] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 885.809702] env[62814]: value = "task-4293864" [ 885.809702] env[62814]: _type = "Task" [ 885.809702] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.821461] env[62814]: INFO nova.compute.manager [-] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Took 1.40 seconds to deallocate network for instance. [ 885.822285] env[62814]: DEBUG nova.compute.provider_tree [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.832520] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293864, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.986451] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5ae2a0f8-4349-4ba4-b586-002803044931 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.758s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 886.239568] env[62814]: DEBUG nova.compute.manager [req-0a3cba99-0d63-41ed-9b98-ee2032c2d6e4 req-fda3397c-da53-4324-96d0-40f3545d5ea6 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Received event network-changed-d9875d9c-cdba-4f00-b6a3-72216a3f9f22 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 886.239769] env[62814]: DEBUG nova.compute.manager [req-0a3cba99-0d63-41ed-9b98-ee2032c2d6e4 req-fda3397c-da53-4324-96d0-40f3545d5ea6 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Refreshing instance network info cache due to event network-changed-d9875d9c-cdba-4f00-b6a3-72216a3f9f22. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 886.240118] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a3cba99-0d63-41ed-9b98-ee2032c2d6e4 req-fda3397c-da53-4324-96d0-40f3545d5ea6 service nova] Acquiring lock "refresh_cache-1e259ec6-d31a-453b-87e0-baa446665d56" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.240294] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a3cba99-0d63-41ed-9b98-ee2032c2d6e4 req-fda3397c-da53-4324-96d0-40f3545d5ea6 service nova] Acquired lock "refresh_cache-1e259ec6-d31a-453b-87e0-baa446665d56" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 886.240556] env[62814]: DEBUG nova.network.neutron [req-0a3cba99-0d63-41ed-9b98-ee2032c2d6e4 req-fda3397c-da53-4324-96d0-40f3545d5ea6 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Refreshing network info cache for port d9875d9c-cdba-4f00-b6a3-72216a3f9f22 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.254413] env[62814]: DEBUG nova.compute.manager [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 886.282678] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 886.282678] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.282678] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 886.282678] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.282678] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 886.282678] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 886.282920] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 886.282920] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 886.282999] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 886.283242] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 886.283416] env[62814]: DEBUG nova.virt.hardware [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 886.284527] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79326e8e-5286-47e9-b7e5-813bcbf71a3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.292935] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca8ccf3-3649-4c03-b213-82015f1c91a2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.319333] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293864, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.327106] env[62814]: DEBUG nova.scheduler.client.report [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 886.341419] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 886.774909] env[62814]: DEBUG nova.network.neutron [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Successfully updated port: 5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.831017] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293864, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.976318} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.834401] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 886.834709] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 886.835637] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.622s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 886.836247] env[62814]: DEBUG nova.compute.manager [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 886.847190] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ca3a19b-45e2-4702-925f-edbed914e0a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.848257] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.952s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 886.848527] env[62814]: DEBUG nova.objects.instance [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lazy-loading 'resources' on Instance uuid 3d3f32fc-276a-49be-b471-01a5d6fc5069 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.855191] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 886.855191] env[62814]: value = "task-4293865" [ 886.855191] env[62814]: _type = "Task" [ 886.855191] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.864544] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293865, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.069452] env[62814]: DEBUG nova.network.neutron [req-0a3cba99-0d63-41ed-9b98-ee2032c2d6e4 req-fda3397c-da53-4324-96d0-40f3545d5ea6 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Updated VIF entry in instance network info cache for port d9875d9c-cdba-4f00-b6a3-72216a3f9f22. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 887.069834] env[62814]: DEBUG nova.network.neutron [req-0a3cba99-0d63-41ed-9b98-ee2032c2d6e4 req-fda3397c-da53-4324-96d0-40f3545d5ea6 service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Updating instance_info_cache with network_info: [{"id": "d9875d9c-cdba-4f00-b6a3-72216a3f9f22", "address": "fa:16:3e:6b:c2:b2", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.250", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd9875d9c-cd", "ovs_interfaceid": "d9875d9c-cdba-4f00-b6a3-72216a3f9f22", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.153038] env[62814]: DEBUG nova.compute.manager [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Received event network-vif-plugged-5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 887.153326] env[62814]: DEBUG oslo_concurrency.lockutils [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] Acquiring lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 887.153514] env[62814]: DEBUG oslo_concurrency.lockutils [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] Lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 887.153665] env[62814]: DEBUG oslo_concurrency.lockutils [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] Lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 887.153801] env[62814]: DEBUG nova.compute.manager [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] No waiting events found dispatching network-vif-plugged-5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 887.153939] env[62814]: WARNING nova.compute.manager [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Received unexpected event network-vif-plugged-5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca for instance with vm_state building and task_state spawning. [ 887.154113] env[62814]: DEBUG nova.compute.manager [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Received event network-changed-5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 887.154269] env[62814]: DEBUG nova.compute.manager [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Refreshing instance network info cache due to event network-changed-5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 887.154449] env[62814]: DEBUG oslo_concurrency.lockutils [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] Acquiring lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.154587] env[62814]: DEBUG oslo_concurrency.lockutils [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] Acquired lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.154740] env[62814]: DEBUG nova.network.neutron [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Refreshing network info cache for port 5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 887.276800] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.352569] env[62814]: DEBUG nova.compute.utils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 887.358573] env[62814]: DEBUG nova.compute.manager [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 887.358573] env[62814]: DEBUG nova.network.neutron [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 887.371971] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293865, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072765} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.372347] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.373468] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a60c2b-87d9-4dc4-adcd-3786a12a4ac5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.405627] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.407691] env[62814]: DEBUG nova.policy [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51348be2493643fb9e5b44f27e5a7a94', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '513135674ee446d19fa8c667a47138a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 887.412694] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44e20a6c-6cc1-4647-b985-bb25ba655bef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.443756] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 887.443756] env[62814]: value = "task-4293866" [ 887.443756] env[62814]: _type = "Task" [ 887.443756] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.454532] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293866, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.573782] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a3cba99-0d63-41ed-9b98-ee2032c2d6e4 req-fda3397c-da53-4324-96d0-40f3545d5ea6 service nova] Releasing lock "refresh_cache-1e259ec6-d31a-453b-87e0-baa446665d56" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 887.703896] env[62814]: DEBUG nova.network.neutron [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 887.832523] env[62814]: DEBUG nova.network.neutron [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.834229] env[62814]: DEBUG nova.network.neutron [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Successfully created port: 742fe6f0-d787-4ca8-be9a-edabfb0646a0 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 887.859130] env[62814]: DEBUG nova.compute.manager [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 887.877588] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb55e93-e513-493b-96ad-534914004cb0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.886086] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b717c2c1-30e7-4bfb-b46f-5e013131585d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.923728] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497704ed-7d7d-4fcf-91c7-bafecc119c40 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.931852] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef58d783-4e6e-42f4-a1df-ea9c9a2b0187 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.945995] env[62814]: DEBUG nova.compute.provider_tree [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.955061] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293866, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.337132] env[62814]: DEBUG oslo_concurrency.lockutils [req-d3f26132-5195-4e35-84ec-108adde952ee req-72e47b82-921e-4526-9dc6-b213001e812e service nova] Releasing lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.337537] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquired lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 888.337706] env[62814]: DEBUG nova.network.neutron [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.451194] env[62814]: DEBUG nova.scheduler.client.report [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 888.458453] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293866, 'name': ReconfigVM_Task, 'duration_secs': 0.993314} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.459927] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.459927] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-227d17a4-c116-4406-90b4-bad59b16ff24 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.465979] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 888.465979] env[62814]: value = "task-4293867" [ 888.465979] env[62814]: _type = "Task" [ 888.465979] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.475806] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293867, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.873046] env[62814]: DEBUG nova.network.neutron [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 888.875815] env[62814]: DEBUG nova.compute.manager [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 888.901415] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 888.901679] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.901864] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 888.902070] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.902234] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 888.902399] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 888.902626] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 888.902800] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 888.902990] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 888.903180] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 888.903361] env[62814]: DEBUG nova.virt.hardware [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 888.904270] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d9dab7-7aba-4b77-8974-72c34c31e0f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.914569] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80cef1d0-505f-4cc6-aed8-69457b434f6c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.959953] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.112s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 888.962342] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.714s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 888.963829] env[62814]: INFO nova.compute.claims [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 888.976269] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293867, 'name': Rename_Task, 'duration_secs': 0.18626} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.976269] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.976516] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d37d2ee3-306a-4356-a899-741d4eeca2b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.984395] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 888.984395] env[62814]: value = "task-4293868" [ 888.984395] env[62814]: _type = "Task" [ 888.984395] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.990357] env[62814]: INFO nova.scheduler.client.report [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Deleted allocations for instance 3d3f32fc-276a-49be-b471-01a5d6fc5069 [ 888.999409] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293868, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.103883] env[62814]: DEBUG nova.network.neutron [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Updating instance_info_cache with network_info: [{"id": "5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca", "address": "fa:16:3e:aa:06:a1", "network": {"id": "4ddbe620-bc9d-4b38-a414-7799c4e7060d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-334880125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef503fd02adc475fa79dda8b379b6514", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c964bdc6-fccc-40d9-bfe2-763b6f05a863", "external-id": "cl2-zone-376", "segmentation_id": 376, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2b0ae5-35", "ovs_interfaceid": "5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.411221] env[62814]: DEBUG nova.compute.manager [req-3b927745-963f-4eac-bdb5-f1b3566be9c8 req-2b8e4c3e-7a45-4851-9d6d-fdd75eee9c2c service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Received event network-vif-plugged-742fe6f0-d787-4ca8-be9a-edabfb0646a0 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 889.411447] env[62814]: DEBUG oslo_concurrency.lockutils [req-3b927745-963f-4eac-bdb5-f1b3566be9c8 req-2b8e4c3e-7a45-4851-9d6d-fdd75eee9c2c service nova] Acquiring lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 889.411659] env[62814]: DEBUG oslo_concurrency.lockutils [req-3b927745-963f-4eac-bdb5-f1b3566be9c8 req-2b8e4c3e-7a45-4851-9d6d-fdd75eee9c2c service nova] Lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 889.411829] env[62814]: DEBUG oslo_concurrency.lockutils [req-3b927745-963f-4eac-bdb5-f1b3566be9c8 req-2b8e4c3e-7a45-4851-9d6d-fdd75eee9c2c service nova] Lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.411995] env[62814]: DEBUG nova.compute.manager [req-3b927745-963f-4eac-bdb5-f1b3566be9c8 req-2b8e4c3e-7a45-4851-9d6d-fdd75eee9c2c service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] No waiting events found dispatching network-vif-plugged-742fe6f0-d787-4ca8-be9a-edabfb0646a0 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 889.414326] env[62814]: WARNING nova.compute.manager [req-3b927745-963f-4eac-bdb5-f1b3566be9c8 req-2b8e4c3e-7a45-4851-9d6d-fdd75eee9c2c service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Received unexpected event network-vif-plugged-742fe6f0-d787-4ca8-be9a-edabfb0646a0 for instance with vm_state building and task_state spawning. [ 889.461233] env[62814]: DEBUG nova.network.neutron [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Successfully updated port: 742fe6f0-d787-4ca8-be9a-edabfb0646a0 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.493855] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293868, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.501787] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4b148f8e-7ccf-4955-b8b0-bf1dd4010d1b tempest-VolumesAssistedSnapshotsTest-475127952 tempest-VolumesAssistedSnapshotsTest-475127952-project-member] Lock "3d3f32fc-276a-49be-b471-01a5d6fc5069" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 39.649s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 889.606112] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Releasing lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 889.607045] env[62814]: DEBUG nova.compute.manager [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Instance network_info: |[{"id": "5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca", "address": "fa:16:3e:aa:06:a1", "network": {"id": "4ddbe620-bc9d-4b38-a414-7799c4e7060d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-334880125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef503fd02adc475fa79dda8b379b6514", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c964bdc6-fccc-40d9-bfe2-763b6f05a863", "external-id": "cl2-zone-376", "segmentation_id": 376, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2b0ae5-35", "ovs_interfaceid": "5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 889.607045] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:06:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c964bdc6-fccc-40d9-bfe2-763b6f05a863', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.614846] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 889.615078] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 889.617697] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7206efd0-7816-420c-8fd8-d15726780d79 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.637892] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.637892] env[62814]: value = "task-4293869" [ 889.637892] env[62814]: _type = "Task" [ 889.637892] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.645849] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293869, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.964835] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-98ba2786-023f-4d36-bdd0-e38bdd23d73c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.964835] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-98ba2786-023f-4d36-bdd0-e38bdd23d73c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 889.964835] env[62814]: DEBUG nova.network.neutron [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 889.998670] env[62814]: DEBUG oslo_vmware.api [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293868, 'name': PowerOnVM_Task, 'duration_secs': 0.816539} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.998754] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.998917] env[62814]: INFO nova.compute.manager [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Took 6.44 seconds to spawn the instance on the hypervisor. [ 889.999134] env[62814]: DEBUG nova.compute.manager [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 890.000454] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-875c0959-e04c-47f9-8413-4bfb9dfc2a49 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.153111] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293869, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.517357] env[62814]: DEBUG nova.network.neutron [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 890.533365] env[62814]: INFO nova.compute.manager [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Took 45.13 seconds to build instance. [ 890.543282] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99bd4193-941c-4498-965f-ce4d08c722c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.557367] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df65fb8-dcc3-4386-aeaa-92ad7c19584e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.593725] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca2ca3e-d9bb-46c1-9db2-037e014c2811 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.604023] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db118338-13e0-4262-8ebb-bd0d86351a13 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.619196] env[62814]: DEBUG nova.compute.provider_tree [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.649906] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293869, 'name': CreateVM_Task, 'duration_secs': 0.519667} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.649906] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 890.649906] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.650102] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 890.650332] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 890.650623] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91d83352-a785-4105-9294-0af249b52262 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.655447] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 890.655447] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b3b79-5abb-e432-5b7e-effbd143d736" [ 890.655447] env[62814]: _type = "Task" [ 890.655447] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.669048] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b3b79-5abb-e432-5b7e-effbd143d736, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.769140] env[62814]: DEBUG nova.network.neutron [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Updating instance_info_cache with network_info: [{"id": "742fe6f0-d787-4ca8-be9a-edabfb0646a0", "address": "fa:16:3e:29:37:79", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap742fe6f0-d7", "ovs_interfaceid": "742fe6f0-d787-4ca8-be9a-edabfb0646a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.962804] env[62814]: INFO nova.compute.manager [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Rebuilding instance [ 891.009390] env[62814]: DEBUG nova.compute.manager [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 891.010656] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70b338f-34f3-4b92-8a7e-9ad70fba862d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.036607] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c736687-ccfb-4394-b7fd-b8fdfdf909fb tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "98884571-8101-4ae9-b2e3-aeed97e3618f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.885s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.125762] env[62814]: DEBUG nova.scheduler.client.report [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 891.169028] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529b3b79-5abb-e432-5b7e-effbd143d736, 'name': SearchDatastore_Task, 'duration_secs': 0.03973} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.169272] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 891.169500] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.169786] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.169972] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.170195] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.170477] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ba06c46-1057-437e-9e96-6bcbf550708a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.182874] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.183068] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.183780] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c38f0120-854d-4cbd-8ba1-14050f77d5c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.189174] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 891.189174] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d01ec5-3bd3-3404-252d-2b8b58914a86" [ 891.189174] env[62814]: _type = "Task" [ 891.189174] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.196774] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d01ec5-3bd3-3404-252d-2b8b58914a86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.271896] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-98ba2786-023f-4d36-bdd0-e38bdd23d73c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 891.272260] env[62814]: DEBUG nova.compute.manager [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Instance network_info: |[{"id": "742fe6f0-d787-4ca8-be9a-edabfb0646a0", "address": "fa:16:3e:29:37:79", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap742fe6f0-d7", "ovs_interfaceid": "742fe6f0-d787-4ca8-be9a-edabfb0646a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 891.272665] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:37:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '742fe6f0-d787-4ca8-be9a-edabfb0646a0', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.280096] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 891.280688] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 891.280894] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9cd7c10f-9851-43f2-b7a5-49a5cd52d92c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.308851] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 891.308851] env[62814]: value = "task-4293870" [ 891.308851] env[62814]: _type = "Task" [ 891.308851] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.317899] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293870, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.445037] env[62814]: DEBUG nova.compute.manager [req-35d15828-fcec-4f12-95fc-bdd3e4e86d79 req-769e57f3-6f55-403f-a13e-5edbd81798b8 service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Received event network-changed-742fe6f0-d787-4ca8-be9a-edabfb0646a0 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 891.445277] env[62814]: DEBUG nova.compute.manager [req-35d15828-fcec-4f12-95fc-bdd3e4e86d79 req-769e57f3-6f55-403f-a13e-5edbd81798b8 service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Refreshing instance network info cache due to event network-changed-742fe6f0-d787-4ca8-be9a-edabfb0646a0. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 891.445453] env[62814]: DEBUG oslo_concurrency.lockutils [req-35d15828-fcec-4f12-95fc-bdd3e4e86d79 req-769e57f3-6f55-403f-a13e-5edbd81798b8 service nova] Acquiring lock "refresh_cache-98ba2786-023f-4d36-bdd0-e38bdd23d73c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.445603] env[62814]: DEBUG oslo_concurrency.lockutils [req-35d15828-fcec-4f12-95fc-bdd3e4e86d79 req-769e57f3-6f55-403f-a13e-5edbd81798b8 service nova] Acquired lock "refresh_cache-98ba2786-023f-4d36-bdd0-e38bdd23d73c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.445761] env[62814]: DEBUG nova.network.neutron [req-35d15828-fcec-4f12-95fc-bdd3e4e86d79 req-769e57f3-6f55-403f-a13e-5edbd81798b8 service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Refreshing network info cache for port 742fe6f0-d787-4ca8-be9a-edabfb0646a0 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.632980] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.671s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 891.633641] env[62814]: DEBUG nova.compute.manager [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 891.636498] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.491s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 891.636719] env[62814]: DEBUG nova.objects.instance [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lazy-loading 'resources' on Instance uuid 448b7bfb-1a90-4c5b-9086-9320cedaf015 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.700194] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d01ec5-3bd3-3404-252d-2b8b58914a86, 'name': SearchDatastore_Task, 'duration_secs': 0.030278} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.703899] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8b91342-4ed0-4089-a3e0-a4e49e2d2ce8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.707804] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 891.707804] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52285d5a-5757-ffcd-4d97-d6d1f431ee06" [ 891.707804] env[62814]: _type = "Task" [ 891.707804] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.716340] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52285d5a-5757-ffcd-4d97-d6d1f431ee06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.822029] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293870, 'name': CreateVM_Task, 'duration_secs': 0.331681} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.822029] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.822029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.822029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.822029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 891.822029] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdc7419e-9382-47c3-9aec-b8526bf3bd80 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.825656] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 891.825656] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb3128-3375-4c15-0200-5a4b83e1b92f" [ 891.825656] env[62814]: _type = "Task" [ 891.825656] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.834109] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb3128-3375-4c15-0200-5a4b83e1b92f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.025480] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 892.026156] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca67a028-512d-4368-b110-31a3e49e16da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.034197] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 892.034197] env[62814]: value = "task-4293871" [ 892.034197] env[62814]: _type = "Task" [ 892.034197] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.047878] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293871, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.139699] env[62814]: DEBUG nova.compute.utils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 892.144169] env[62814]: DEBUG nova.compute.manager [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Not allocating networking since 'none' was specified. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 892.206455] env[62814]: DEBUG nova.network.neutron [req-35d15828-fcec-4f12-95fc-bdd3e4e86d79 req-769e57f3-6f55-403f-a13e-5edbd81798b8 service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Updated VIF entry in instance network info cache for port 742fe6f0-d787-4ca8-be9a-edabfb0646a0. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.208021] env[62814]: DEBUG nova.network.neutron [req-35d15828-fcec-4f12-95fc-bdd3e4e86d79 req-769e57f3-6f55-403f-a13e-5edbd81798b8 service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Updating instance_info_cache with network_info: [{"id": "742fe6f0-d787-4ca8-be9a-edabfb0646a0", "address": "fa:16:3e:29:37:79", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap742fe6f0-d7", "ovs_interfaceid": "742fe6f0-d787-4ca8-be9a-edabfb0646a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.218475] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52285d5a-5757-ffcd-4d97-d6d1f431ee06, 'name': SearchDatastore_Task, 'duration_secs': 0.040617} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.221761] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.222044] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 310e22c3-ff0a-4e6b-aa46-8812fa014dfa/310e22c3-ff0a-4e6b-aa46-8812fa014dfa.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.222851] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-296d302e-8263-4497-bb8d-55bd187c48e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.232454] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 892.232454] env[62814]: value = "task-4293872" [ 892.232454] env[62814]: _type = "Task" [ 892.232454] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.244388] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293872, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.341254] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb3128-3375-4c15-0200-5a4b83e1b92f, 'name': SearchDatastore_Task, 'duration_secs': 0.039107} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.343915] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.344185] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 892.344425] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.344638] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 892.344846] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 892.345308] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-520b1989-cbb1-4016-828e-e74e32cf7191 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.365222] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 892.365808] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 892.366656] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfdaf38d-5d12-4729-b92d-c3d6fdad13c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.375626] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 892.375626] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5237d89f-0e8b-84c3-d8fc-e44e04254efb" [ 892.375626] env[62814]: _type = "Task" [ 892.375626] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.384442] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5237d89f-0e8b-84c3-d8fc-e44e04254efb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.546830] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293871, 'name': PowerOffVM_Task, 'duration_secs': 0.23402} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.547124] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 892.547408] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 892.548240] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10461a3-0ceb-42bf-a5d5-e3c9e68369db {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.557454] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 892.557696] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9919cdda-95f3-4676-8542-d030c9a890d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.582648] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 892.582877] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 892.583079] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Deleting the datastore file [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.583348] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7787d59-1934-44e3-adab-41edad782818 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.594632] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 892.594632] env[62814]: value = "task-4293874" [ 892.594632] env[62814]: _type = "Task" [ 892.594632] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.606036] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293874, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.645461] env[62814]: DEBUG nova.compute.manager [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 892.648178] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d01f90-9f1a-4ced-ac23-880f7b6a5173 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.656861] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab99a3a7-2dd4-440d-8273-d99ac588cf60 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.698851] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e71c2d-bbaa-4bea-a7e9-6712178551cc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.707544] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e537f09d-0556-4c9f-bf8c-15c24f1b5248 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.712809] env[62814]: DEBUG oslo_concurrency.lockutils [req-35d15828-fcec-4f12-95fc-bdd3e4e86d79 req-769e57f3-6f55-403f-a13e-5edbd81798b8 service nova] Releasing lock "refresh_cache-98ba2786-023f-4d36-bdd0-e38bdd23d73c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.723864] env[62814]: DEBUG nova.compute.provider_tree [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.751393] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293872, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.894722] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5237d89f-0e8b-84c3-d8fc-e44e04254efb, 'name': SearchDatastore_Task, 'duration_secs': 0.041284} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.895413] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27179b1c-ea3e-4ad5-934a-63bddfa2ced8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.903785] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 892.903785] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b46de6-e23e-a03f-3f97-a3e2b84fd989" [ 892.903785] env[62814]: _type = "Task" [ 892.903785] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.920903] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b46de6-e23e-a03f-3f97-a3e2b84fd989, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.105981] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293874, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.381285} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.105981] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.105981] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 893.105981] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 893.226984] env[62814]: DEBUG nova.scheduler.client.report [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 893.245021] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293872, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.696604} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.246023] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 310e22c3-ff0a-4e6b-aa46-8812fa014dfa/310e22c3-ff0a-4e6b-aa46-8812fa014dfa.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.246245] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.246501] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-004b45cc-fdac-4eed-9f9f-6b612a72228e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.253612] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 893.253612] env[62814]: value = "task-4293875" [ 893.253612] env[62814]: _type = "Task" [ 893.253612] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.261414] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293875, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.414858] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b46de6-e23e-a03f-3f97-a3e2b84fd989, 'name': SearchDatastore_Task, 'duration_secs': 0.018888} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.415159] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 893.415432] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98ba2786-023f-4d36-bdd0-e38bdd23d73c/98ba2786-023f-4d36-bdd0-e38bdd23d73c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 893.415689] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2006b5ef-a870-4903-8f8d-7408cb74159c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.422530] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 893.422530] env[62814]: value = "task-4293876" [ 893.422530] env[62814]: _type = "Task" [ 893.422530] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.430931] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293876, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.657496] env[62814]: DEBUG nova.compute.manager [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 893.683233] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 893.683547] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 893.683660] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 893.683913] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 893.684083] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 893.684255] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 893.684468] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 893.684783] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 893.684836] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 893.684952] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 893.685136] env[62814]: DEBUG nova.virt.hardware [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 893.686262] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3347692e-2c2b-4e78-99a0-85bfb0b19c7c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.694532] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bcb311-0975-4343-af8f-1024ffe690d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.709553] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 893.715267] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Creating folder: Project (d6afed91240446e681ad907812971ed6). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 893.715659] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1eebec0-8cb7-422b-bf81-8af1dcba4585 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.725472] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Created folder: Project (d6afed91240446e681ad907812971ed6) in parent group-v845547. [ 893.725654] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Creating folder: Instances. Parent ref: group-v845722. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 893.725872] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5e1d340d-721d-4ea6-9552-21abc515fdba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.731333] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.095s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 893.734029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.854s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 893.737068] env[62814]: INFO nova.compute.claims [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.739158] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Created folder: Instances in parent group-v845722. [ 893.739479] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 893.739742] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 893.740046] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-efe1d8f6-32a9-4ee1-a2dc-b17359ad60fb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.759696] env[62814]: INFO nova.scheduler.client.report [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Deleted allocations for instance 448b7bfb-1a90-4c5b-9086-9320cedaf015 [ 893.760598] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 893.760598] env[62814]: value = "task-4293879" [ 893.760598] env[62814]: _type = "Task" [ 893.760598] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.770037] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293875, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065831} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.770507] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.771333] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bca5eef-02e8-4453-bde0-637574aeea50 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.776951] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293879, 'name': CreateVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.801050] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 310e22c3-ff0a-4e6b-aa46-8812fa014dfa/310e22c3-ff0a-4e6b-aa46-8812fa014dfa.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.801857] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b44b9e6-060c-4cf8-9ac1-89df48d159c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.827027] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 893.827027] env[62814]: value = "task-4293880" [ 893.827027] env[62814]: _type = "Task" [ 893.827027] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.836975] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293880, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.940702] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293876, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.147291] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 894.147291] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.147291] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 894.147291] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.147291] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 894.147291] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 894.147291] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 894.147291] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 894.147291] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 894.147291] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 894.148557] env[62814]: DEBUG nova.virt.hardware [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 894.149787] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e63e006-8c86-432d-b61d-e978fe85d9ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.158251] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df649507-9023-419e-b3cb-088fb0b25784 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.173858] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.181074] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 894.181074] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 894.181074] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f54c9dde-b843-4750-a144-24dce278b439 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.197227] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.197227] env[62814]: value = "task-4293881" [ 894.197227] env[62814]: _type = "Task" [ 894.197227] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.208557] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293881, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.274068] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b56e35c-e700-419f-8b18-54e1420559f3 tempest-ServerGroupTestJSON-1526461688 tempest-ServerGroupTestJSON-1526461688-project-member] Lock "448b7bfb-1a90-4c5b-9086-9320cedaf015" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.053s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 894.284027] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293879, 'name': CreateVM_Task, 'duration_secs': 0.464688} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.284027] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 894.284027] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.284027] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.284027] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 894.284379] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec745cd5-4bf3-4f83-a5e1-b4229142fd2c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.290290] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 894.290290] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c39fdc-69f1-371d-9fc6-58bda0f8aa15" [ 894.290290] env[62814]: _type = "Task" [ 894.290290] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.299186] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c39fdc-69f1-371d-9fc6-58bda0f8aa15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.335906] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293880, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.434172] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293876, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.75852} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.437200] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98ba2786-023f-4d36-bdd0-e38bdd23d73c/98ba2786-023f-4d36-bdd0-e38bdd23d73c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 894.437200] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 894.437200] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7d7efe7b-3a06-44ef-9b27-5256b08102a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.443547] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 894.443547] env[62814]: value = "task-4293882" [ 894.443547] env[62814]: _type = "Task" [ 894.443547] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.453177] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293882, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.707823] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293881, 'name': CreateVM_Task, 'duration_secs': 0.300596} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.708185] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 894.708729] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.802645] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c39fdc-69f1-371d-9fc6-58bda0f8aa15, 'name': SearchDatastore_Task, 'duration_secs': 0.009661} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.805994] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 894.805994] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.805994] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.806128] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.806273] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 894.806767] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.807061] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 894.807723] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ddb2c7f-5c9e-48eb-bfec-f11ba502a5ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.809114] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0debef8c-2add-4980-a91b-8931284d6e1c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.815586] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 894.815586] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526d84c0-ed3e-e8e3-ceb2-51190db59deb" [ 894.815586] env[62814]: _type = "Task" [ 894.815586] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.821821] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 894.822078] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 894.825605] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5f5857c-c898-4106-b38c-47bcdb4df9fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.831872] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526d84c0-ed3e-e8e3-ceb2-51190db59deb, 'name': SearchDatastore_Task, 'duration_secs': 0.00993} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.835253] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 894.835491] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 894.835700] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.837660] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 894.837660] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529d7a0c-2e00-a738-f3ab-3997d82e9111" [ 894.837660] env[62814]: _type = "Task" [ 894.837660] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.840987] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293880, 'name': ReconfigVM_Task, 'duration_secs': 0.539478} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.844336] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 310e22c3-ff0a-4e6b-aa46-8812fa014dfa/310e22c3-ff0a-4e6b-aa46-8812fa014dfa.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.847264] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f9f70143-0dfe-4d65-a715-b909c447c899 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.857565] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 894.857565] env[62814]: value = "task-4293883" [ 894.857565] env[62814]: _type = "Task" [ 894.857565] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.857807] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529d7a0c-2e00-a738-f3ab-3997d82e9111, 'name': SearchDatastore_Task, 'duration_secs': 0.008942} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.862500] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b78501d-a8e7-450c-98d0-97c4729ec8cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.873349] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293883, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.874788] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 894.874788] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527bc21a-5a36-0d8f-3b51-d332235443d4" [ 894.874788] env[62814]: _type = "Task" [ 894.874788] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.882961] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527bc21a-5a36-0d8f-3b51-d332235443d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.956593] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293882, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063406} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.956593] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.957356] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3d2581-ed16-43b7-b4b0-37a97ee71dc3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.981176] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 98ba2786-023f-4d36-bdd0-e38bdd23d73c/98ba2786-023f-4d36-bdd0-e38bdd23d73c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.984405] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8da0370e-0113-4390-8c86-4a6d0d0c69ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.004876] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 895.004876] env[62814]: value = "task-4293884" [ 895.004876] env[62814]: _type = "Task" [ 895.004876] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.017237] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293884, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.329925] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a277ca6d-ce85-421f-8552-d934a20c61dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.337776] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45fb5fd-4945-4e15-a5cd-300070881cd7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.373410] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6efef64-173c-4981-9cb0-a53ebe8a45ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.386534] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293883, 'name': Rename_Task, 'duration_secs': 0.136653} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.389160] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.389513] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9316528e-2a14-42b4-b5e5-7a1677a08ae3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.391981] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c3f9de8-5ee4-45dc-ad4a-043193c01115 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.399196] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527bc21a-5a36-0d8f-3b51-d332235443d4, 'name': SearchDatastore_Task, 'duration_secs': 0.009524} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.400191] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 895.400557] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d7332882-19b3-4ab3-8ea6-51d33c584844/d7332882-19b3-4ab3-8ea6-51d33c584844.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 895.400894] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 895.402152] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.402152] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68623f86-477b-47c7-9669-ad1ab07aa17e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.411534] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d6054b1-5a08-4175-b31c-ab7e076b5e62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.413799] env[62814]: DEBUG nova.compute.provider_tree [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 895.417554] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 895.417554] env[62814]: value = "task-4293885" [ 895.417554] env[62814]: _type = "Task" [ 895.417554] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.421587] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 895.421587] env[62814]: value = "task-4293886" [ 895.421587] env[62814]: _type = "Task" [ 895.421587] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.429363] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.429595] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.430350] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293885, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.431574] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cb0b77f-6615-4879-9bd4-f7133b0992e1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.438104] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.440404] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 895.440404] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef1cc4-c6c1-ac93-6b9c-b3d2be2c6606" [ 895.440404] env[62814]: _type = "Task" [ 895.440404] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.448739] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef1cc4-c6c1-ac93-6b9c-b3d2be2c6606, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.516863] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293884, 'name': ReconfigVM_Task, 'duration_secs': 0.268218} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.516863] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 98ba2786-023f-4d36-bdd0-e38bdd23d73c/98ba2786-023f-4d36-bdd0-e38bdd23d73c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.517363] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7fd5db4b-a02a-4b8f-bfe3-8398205b9269 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.529780] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 895.529780] env[62814]: value = "task-4293887" [ 895.529780] env[62814]: _type = "Task" [ 895.529780] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.538736] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293887, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.919572] env[62814]: DEBUG nova.scheduler.client.report [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 895.936664] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293886, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.939888] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293885, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.950175] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef1cc4-c6c1-ac93-6b9c-b3d2be2c6606, 'name': SearchDatastore_Task, 'duration_secs': 0.00961} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.951115] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-317b0158-b38e-4cdc-aeee-0043f972fd75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.959118] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 895.959118] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5222193c-09ad-ceeb-7917-3db153efb5db" [ 895.959118] env[62814]: _type = "Task" [ 895.959118] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.970891] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5222193c-09ad-ceeb-7917-3db153efb5db, 'name': SearchDatastore_Task, 'duration_secs': 0.00932} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.971130] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 895.971398] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 895.971686] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5df5bcd6-0a8f-4df8-b1f1-82b81005630c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.979539] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 895.979539] env[62814]: value = "task-4293888" [ 895.979539] env[62814]: _type = "Task" [ 895.979539] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.991414] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293888, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.039962] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293887, 'name': Rename_Task, 'duration_secs': 0.24655} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.040254] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.041114] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49c2a122-3891-4c8c-b3b8-737c1cd306c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.048188] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 896.048188] env[62814]: value = "task-4293889" [ 896.048188] env[62814]: _type = "Task" [ 896.048188] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.058204] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293889, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.432811] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.699s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 896.433180] env[62814]: DEBUG nova.compute.manager [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 896.436406] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.495s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 896.436641] env[62814]: DEBUG nova.objects.instance [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lazy-loading 'resources' on Instance uuid 17bfe703-ff96-4cb7-8535-eed57e10d673 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.443045] env[62814]: DEBUG oslo_vmware.api [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293885, 'name': PowerOnVM_Task, 'duration_secs': 0.904134} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.447275] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.447637] env[62814]: INFO nova.compute.manager [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Took 10.19 seconds to spawn the instance on the hypervisor. [ 896.447893] env[62814]: DEBUG nova.compute.manager [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 896.448317] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293886, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.529929} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.449186] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d04d835-2645-4a63-9b85-4536a2d40fbb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.452147] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d7332882-19b3-4ab3-8ea6-51d33c584844/d7332882-19b3-4ab3-8ea6-51d33c584844.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 896.452389] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 896.453515] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b672d5e-82c5-48a6-9745-d1d52f68061e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.460845] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 896.460845] env[62814]: value = "task-4293890" [ 896.460845] env[62814]: _type = "Task" [ 896.460845] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.476117] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293890, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.493152] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293888, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.558521] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293889, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.940574] env[62814]: DEBUG nova.compute.utils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 896.945380] env[62814]: DEBUG nova.compute.manager [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 896.945380] env[62814]: DEBUG nova.network.neutron [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 896.980869] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293890, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.111057} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.986274] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 896.987248] env[62814]: INFO nova.compute.manager [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Took 46.97 seconds to build instance. [ 896.988762] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99c56ff-f6cf-4e0c-bfcc-1d31fdc64efb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.003089] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293888, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535098} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.015470] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 897.016122] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.026771] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] d7332882-19b3-4ab3-8ea6-51d33c584844/d7332882-19b3-4ab3-8ea6-51d33c584844.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.030846] env[62814]: DEBUG nova.policy [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '58359539fd4545cdbf200ec364e43834', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dedbeff832a4ac48b0aa01bd6acc3f1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 897.032329] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6339bd64-f6c8-4f90-955c-315289571f12 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.034526] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6618c1de-7a74-4aa0-89d4-6377856fb619 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.059578] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 897.059578] env[62814]: value = "task-4293891" [ 897.059578] env[62814]: _type = "Task" [ 897.059578] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.061709] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 897.061709] env[62814]: value = "task-4293892" [ 897.061709] env[62814]: _type = "Task" [ 897.061709] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.074717] env[62814]: DEBUG oslo_vmware.api [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4293889, 'name': PowerOnVM_Task, 'duration_secs': 0.788179} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.075583] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.075810] env[62814]: INFO nova.compute.manager [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Took 8.20 seconds to spawn the instance on the hypervisor. [ 897.076051] env[62814]: DEBUG nova.compute.manager [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 897.077179] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00222e0c-773c-454a-8c63-bde37c5b95c2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.085835] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293892, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.086124] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293891, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.433311] env[62814]: DEBUG nova.network.neutron [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Successfully created port: b8613625-0ec7-47e4-b3d3-8a9bef4be8c7 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 897.446415] env[62814]: DEBUG nova.compute.manager [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 897.500292] env[62814]: DEBUG oslo_concurrency.lockutils [None req-763057fb-5f4e-4f10-b96f-a5d61c11cc41 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.492s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 897.515174] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79768ccf-274e-4aaa-aa09-677de42b2987 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.524672] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8907a906-3204-4c77-ba65-1e69ccb60a0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.558417] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6917447c-15ed-4bc5-a9e2-a0e024d15097 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.571154] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decaeecd-df80-4a51-848f-4b3cf46acf4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.580294] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293891, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.232541} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.581349] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 897.582255] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb00dab-dcef-4907-ae52-7ae44031147d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.596054] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293892, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.600812] env[62814]: DEBUG nova.compute.provider_tree [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.604867] env[62814]: INFO nova.compute.manager [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Took 44.11 seconds to build instance. [ 897.621302] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.622469] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed72ce60-631d-4131-a35a-cceaa5b2081d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.644106] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 897.644106] env[62814]: value = "task-4293893" [ 897.644106] env[62814]: _type = "Task" [ 897.644106] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.653157] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293893, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.078527] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293892, 'name': ReconfigVM_Task, 'duration_secs': 0.752388} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.078891] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Reconfigured VM instance instance-00000040 to attach disk [datastore2] d7332882-19b3-4ab3-8ea6-51d33c584844/d7332882-19b3-4ab3-8ea6-51d33c584844.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.079705] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-40ecff8e-8947-4eb6-a11c-f8563c86346b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.086615] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 898.086615] env[62814]: value = "task-4293894" [ 898.086615] env[62814]: _type = "Task" [ 898.086615] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.096929] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293894, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.106235] env[62814]: DEBUG nova.scheduler.client.report [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 898.122868] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f7df019-eecb-4e3e-a7e5-3f31457671cf tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.036s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.158140] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293893, 'name': ReconfigVM_Task, 'duration_secs': 0.315099} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.158498] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.159658] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a93547d-0c33-43ce-880c-0467f442739e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.168920] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 898.168920] env[62814]: value = "task-4293895" [ 898.168920] env[62814]: _type = "Task" [ 898.168920] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.178597] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293895, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.205876] env[62814]: DEBUG nova.compute.manager [req-f98dd5c4-3e0f-47e8-9efe-c10d7ee32108 req-8f45bab1-5411-498e-98f6-d8401f8678e0 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Received event network-changed-5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 898.205876] env[62814]: DEBUG nova.compute.manager [req-f98dd5c4-3e0f-47e8-9efe-c10d7ee32108 req-8f45bab1-5411-498e-98f6-d8401f8678e0 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Refreshing instance network info cache due to event network-changed-5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 898.205876] env[62814]: DEBUG oslo_concurrency.lockutils [req-f98dd5c4-3e0f-47e8-9efe-c10d7ee32108 req-8f45bab1-5411-498e-98f6-d8401f8678e0 service nova] Acquiring lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.205876] env[62814]: DEBUG oslo_concurrency.lockutils [req-f98dd5c4-3e0f-47e8-9efe-c10d7ee32108 req-8f45bab1-5411-498e-98f6-d8401f8678e0 service nova] Acquired lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 898.205876] env[62814]: DEBUG nova.network.neutron [req-f98dd5c4-3e0f-47e8-9efe-c10d7ee32108 req-8f45bab1-5411-498e-98f6-d8401f8678e0 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Refreshing network info cache for port 5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 898.457244] env[62814]: DEBUG nova.compute.manager [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 898.486594] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='21129ff59ca770d1f0c570836b142553',container_format='bare',created_at=2025-06-21T05:38:14Z,direct_url=,disk_format='vmdk',id=577d3b83-11e9-430b-a3da-3fc2f382fd79,min_disk=1,min_ram=0,name='tempest-test-snap-1684845904',owner='3dedbeff832a4ac48b0aa01bd6acc3f1',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2025-06-21T05:38:30Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 898.486855] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.487015] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 898.487205] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.487351] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 898.487497] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 898.487719] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 898.487874] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 898.488948] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 898.489353] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 898.489421] env[62814]: DEBUG nova.virt.hardware [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 898.490347] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985b342a-377b-4f94-9060-949ac0f6e55f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.499815] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41073b55-ce10-4ef9-a5f3-d3a2016da2b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.603092] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293894, 'name': Rename_Task, 'duration_secs': 0.167914} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.603092] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 898.603092] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd7e6edc-581b-44f2-8452-66882f877f5e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.609119] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 898.609119] env[62814]: value = "task-4293896" [ 898.609119] env[62814]: _type = "Task" [ 898.609119] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.612977] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.177s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.615070] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.254s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.616530] env[62814]: INFO nova.compute.claims [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.625157] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293896, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.658951] env[62814]: INFO nova.scheduler.client.report [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Deleted allocations for instance 17bfe703-ff96-4cb7-8535-eed57e10d673 [ 898.681096] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293895, 'name': Rename_Task, 'duration_secs': 0.180979} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.681393] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 898.681683] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9c7347c5-e99b-4c24-ba11-1c7f9925c2f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.688772] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 898.688772] env[62814]: value = "task-4293897" [ 898.688772] env[62814]: _type = "Task" [ 898.688772] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.697539] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293897, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.864048] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 898.864323] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.864556] env[62814]: INFO nova.compute.manager [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Rebooting instance [ 899.119814] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293896, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.142441] env[62814]: DEBUG nova.network.neutron [req-f98dd5c4-3e0f-47e8-9efe-c10d7ee32108 req-8f45bab1-5411-498e-98f6-d8401f8678e0 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Updated VIF entry in instance network info cache for port 5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 899.142843] env[62814]: DEBUG nova.network.neutron [req-f98dd5c4-3e0f-47e8-9efe-c10d7ee32108 req-8f45bab1-5411-498e-98f6-d8401f8678e0 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Updating instance_info_cache with network_info: [{"id": "5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca", "address": "fa:16:3e:aa:06:a1", "network": {"id": "4ddbe620-bc9d-4b38-a414-7799c4e7060d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-334880125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef503fd02adc475fa79dda8b379b6514", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c964bdc6-fccc-40d9-bfe2-763b6f05a863", "external-id": "cl2-zone-376", "segmentation_id": 376, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2b0ae5-35", "ovs_interfaceid": "5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.168743] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51a29481-dfb9-4e58-accc-fe1403e70dcb tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "17bfe703-ff96-4cb7-8535-eed57e10d673" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.272s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.200974] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293897, 'name': PowerOnVM_Task} progress is 79%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.259897] env[62814]: DEBUG nova.compute.manager [req-538600c1-ef16-4d81-a69e-8f56b87d6c90 req-66b94af2-2274-4af8-867b-035393e16a4c service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Received event network-changed-742fe6f0-d787-4ca8-be9a-edabfb0646a0 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 899.259897] env[62814]: DEBUG nova.compute.manager [req-538600c1-ef16-4d81-a69e-8f56b87d6c90 req-66b94af2-2274-4af8-867b-035393e16a4c service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Refreshing instance network info cache due to event network-changed-742fe6f0-d787-4ca8-be9a-edabfb0646a0. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 899.260132] env[62814]: DEBUG oslo_concurrency.lockutils [req-538600c1-ef16-4d81-a69e-8f56b87d6c90 req-66b94af2-2274-4af8-867b-035393e16a4c service nova] Acquiring lock "refresh_cache-98ba2786-023f-4d36-bdd0-e38bdd23d73c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.260290] env[62814]: DEBUG oslo_concurrency.lockutils [req-538600c1-ef16-4d81-a69e-8f56b87d6c90 req-66b94af2-2274-4af8-867b-035393e16a4c service nova] Acquired lock "refresh_cache-98ba2786-023f-4d36-bdd0-e38bdd23d73c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.260455] env[62814]: DEBUG nova.network.neutron [req-538600c1-ef16-4d81-a69e-8f56b87d6c90 req-66b94af2-2274-4af8-867b-035393e16a4c service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Refreshing network info cache for port 742fe6f0-d787-4ca8-be9a-edabfb0646a0 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 899.299064] env[62814]: DEBUG nova.network.neutron [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Successfully updated port: b8613625-0ec7-47e4-b3d3-8a9bef4be8c7 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 899.387274] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.631171] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293896, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.647395] env[62814]: DEBUG oslo_concurrency.lockutils [req-f98dd5c4-3e0f-47e8-9efe-c10d7ee32108 req-8f45bab1-5411-498e-98f6-d8401f8678e0 service nova] Releasing lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 899.647395] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquired lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.647395] env[62814]: DEBUG nova.network.neutron [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 899.702166] env[62814]: DEBUG oslo_vmware.api [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293897, 'name': PowerOnVM_Task, 'duration_secs': 0.893673} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.702166] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 899.702166] env[62814]: DEBUG nova.compute.manager [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 899.703159] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae91b59-d071-4cb5-8d06-f2e8052c8a16 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.802088] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "refresh_cache-af6ef867-dc9a-4db6-8582-a6198e5caa77" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.802265] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "refresh_cache-af6ef867-dc9a-4db6-8582-a6198e5caa77" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.802458] env[62814]: DEBUG nova.network.neutron [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.105629] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e185bab1-eed7-4a04-b76c-49f5ce21c52d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.117153] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af0158a-72e1-4fbb-8259-6bdee23598a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.125695] env[62814]: DEBUG oslo_vmware.api [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293896, 'name': PowerOnVM_Task, 'duration_secs': 1.061833} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.150525] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.150765] env[62814]: INFO nova.compute.manager [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Took 6.49 seconds to spawn the instance on the hypervisor. [ 900.150950] env[62814]: DEBUG nova.compute.manager [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 900.153823] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5635e7b5-f5af-45a0-9def-d8f77ef0ec28 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.156639] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b57286-11bf-46a3-97a5-702ecfb6b056 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.171320] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1465f97f-8f0d-4c0f-93bf-b17080a16728 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.184911] env[62814]: DEBUG nova.compute.provider_tree [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.221507] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 900.295573] env[62814]: DEBUG nova.network.neutron [req-538600c1-ef16-4d81-a69e-8f56b87d6c90 req-66b94af2-2274-4af8-867b-035393e16a4c service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Updated VIF entry in instance network info cache for port 742fe6f0-d787-4ca8-be9a-edabfb0646a0. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 900.295573] env[62814]: DEBUG nova.network.neutron [req-538600c1-ef16-4d81-a69e-8f56b87d6c90 req-66b94af2-2274-4af8-867b-035393e16a4c service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Updating instance_info_cache with network_info: [{"id": "742fe6f0-d787-4ca8-be9a-edabfb0646a0", "address": "fa:16:3e:29:37:79", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap742fe6f0-d7", "ovs_interfaceid": "742fe6f0-d787-4ca8-be9a-edabfb0646a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.357167] env[62814]: DEBUG nova.compute.manager [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Received event network-vif-plugged-b8613625-0ec7-47e4-b3d3-8a9bef4be8c7 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 900.357167] env[62814]: DEBUG oslo_concurrency.lockutils [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] Acquiring lock "af6ef867-dc9a-4db6-8582-a6198e5caa77-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 900.357167] env[62814]: DEBUG oslo_concurrency.lockutils [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] Lock "af6ef867-dc9a-4db6-8582-a6198e5caa77-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.357167] env[62814]: DEBUG oslo_concurrency.lockutils [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] Lock "af6ef867-dc9a-4db6-8582-a6198e5caa77-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.357167] env[62814]: DEBUG nova.compute.manager [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] No waiting events found dispatching network-vif-plugged-b8613625-0ec7-47e4-b3d3-8a9bef4be8c7 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 900.357167] env[62814]: WARNING nova.compute.manager [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Received unexpected event network-vif-plugged-b8613625-0ec7-47e4-b3d3-8a9bef4be8c7 for instance with vm_state building and task_state spawning. [ 900.357958] env[62814]: DEBUG nova.compute.manager [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Received event network-changed-b8613625-0ec7-47e4-b3d3-8a9bef4be8c7 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 900.358404] env[62814]: DEBUG nova.compute.manager [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Refreshing instance network info cache due to event network-changed-b8613625-0ec7-47e4-b3d3-8a9bef4be8c7. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 900.358837] env[62814]: DEBUG oslo_concurrency.lockutils [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] Acquiring lock "refresh_cache-af6ef867-dc9a-4db6-8582-a6198e5caa77" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.414122] env[62814]: DEBUG nova.network.neutron [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 900.681752] env[62814]: INFO nova.compute.manager [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Took 43.45 seconds to build instance. [ 900.687618] env[62814]: DEBUG nova.scheduler.client.report [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 900.798990] env[62814]: DEBUG oslo_concurrency.lockutils [req-538600c1-ef16-4d81-a69e-8f56b87d6c90 req-66b94af2-2274-4af8-867b-035393e16a4c service nova] Releasing lock "refresh_cache-98ba2786-023f-4d36-bdd0-e38bdd23d73c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 900.821589] env[62814]: DEBUG nova.network.neutron [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Updating instance_info_cache with network_info: [{"id": "5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca", "address": "fa:16:3e:aa:06:a1", "network": {"id": "4ddbe620-bc9d-4b38-a414-7799c4e7060d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-334880125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef503fd02adc475fa79dda8b379b6514", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c964bdc6-fccc-40d9-bfe2-763b6f05a863", "external-id": "cl2-zone-376", "segmentation_id": 376, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2b0ae5-35", "ovs_interfaceid": "5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.831179] env[62814]: DEBUG nova.network.neutron [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Updating instance_info_cache with network_info: [{"id": "b8613625-0ec7-47e4-b3d3-8a9bef4be8c7", "address": "fa:16:3e:f3:84:b8", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8613625-0e", "ovs_interfaceid": "b8613625-0ec7-47e4-b3d3-8a9bef4be8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.185725] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8df212d5-3ee5-4e11-b268-a2cab25338be tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "d7332882-19b3-4ab3-8ea6-51d33c584844" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.013s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.197846] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.583s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 901.198518] env[62814]: DEBUG nova.compute.manager [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 901.204974] env[62814]: DEBUG oslo_concurrency.lockutils [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 30.772s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.204974] env[62814]: DEBUG nova.objects.instance [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 901.326905] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Releasing lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.332569] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "5c574786-64a4-4f07-a267-101ecaaa6938" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.332811] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "5c574786-64a4-4f07-a267-101ecaaa6938" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.334249] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "refresh_cache-af6ef867-dc9a-4db6-8582-a6198e5caa77" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.334782] env[62814]: DEBUG nova.compute.manager [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Instance network_info: |[{"id": "b8613625-0ec7-47e4-b3d3-8a9bef4be8c7", "address": "fa:16:3e:f3:84:b8", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8613625-0e", "ovs_interfaceid": "b8613625-0ec7-47e4-b3d3-8a9bef4be8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 901.335624] env[62814]: DEBUG oslo_concurrency.lockutils [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] Acquired lock "refresh_cache-af6ef867-dc9a-4db6-8582-a6198e5caa77" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 901.335624] env[62814]: DEBUG nova.network.neutron [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Refreshing network info cache for port b8613625-0ec7-47e4-b3d3-8a9bef4be8c7 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 901.338857] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:84:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3db2ab9e-1244-4377-b05f-ab76003f2428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b8613625-0ec7-47e4-b3d3-8a9bef4be8c7', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.347028] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 901.351046] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.351625] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-34ffc9bd-9f50-4c2d-9827-2e9a50a1c6af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.373560] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.373560] env[62814]: value = "task-4293898" [ 901.373560] env[62814]: _type = "Task" [ 901.373560] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.382615] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293898, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.704404] env[62814]: DEBUG nova.compute.utils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 901.706293] env[62814]: DEBUG nova.compute.manager [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 901.706597] env[62814]: DEBUG nova.network.neutron [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 901.803881] env[62814]: DEBUG nova.policy [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd71116c0cfd4b509e4070ecc84613f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd82993ef7dfa4d2f8f39db3577dd321a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 901.820045] env[62814]: DEBUG nova.network.neutron [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Updated VIF entry in instance network info cache for port b8613625-0ec7-47e4-b3d3-8a9bef4be8c7. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.822322] env[62814]: DEBUG nova.network.neutron [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Updating instance_info_cache with network_info: [{"id": "b8613625-0ec7-47e4-b3d3-8a9bef4be8c7", "address": "fa:16:3e:f3:84:b8", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb8613625-0e", "ovs_interfaceid": "b8613625-0ec7-47e4-b3d3-8a9bef4be8c7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.833305] env[62814]: INFO nova.compute.manager [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Rebuilding instance [ 901.836608] env[62814]: DEBUG nova.compute.manager [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 901.838983] env[62814]: DEBUG nova.compute.manager [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 901.843197] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e386b174-c7ed-469a-a5a6-a647948b88cc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.888030] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293898, 'name': CreateVM_Task, 'duration_secs': 0.501203} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.892615] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.893047] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.893819] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "[datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 901.893819] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 901.893942] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af41bd90-96e0-4be2-b1a6-c0af30a54526 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.898309] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 901.898309] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d46dba-1cd8-4795-35b5-0f49b2882d36" [ 901.898309] env[62814]: _type = "Task" [ 901.898309] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.906487] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d46dba-1cd8-4795-35b5-0f49b2882d36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.913217] env[62814]: DEBUG nova.compute.manager [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 901.914050] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df29291a-d51d-45f2-9bce-7978fb33305d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.931460] env[62814]: INFO nova.compute.manager [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Rebuilding instance [ 901.992945] env[62814]: DEBUG nova.compute.manager [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 901.995878] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d879870-f7b7-47d7-ae7a-5bb4bca576bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.210127] env[62814]: DEBUG nova.compute.manager [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 902.214278] env[62814]: DEBUG oslo_concurrency.lockutils [None req-008eb800-753b-421b-acdf-07f2293defdd tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 902.216022] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 29.998s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 902.324099] env[62814]: DEBUG oslo_concurrency.lockutils [req-865ddaad-05fc-4f0c-bb21-6a1319932713 req-fdb14f09-5eb5-41b6-8535-e2e2f20426d4 service nova] Releasing lock "refresh_cache-af6ef867-dc9a-4db6-8582-a6198e5caa77" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 902.367433] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 902.413740] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "[datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 902.414096] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Processing image 577d3b83-11e9-430b-a3da-3fc2f382fd79 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 902.414413] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79/577d3b83-11e9-430b-a3da-3fc2f382fd79.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.414617] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "[datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79/577d3b83-11e9-430b-a3da-3fc2f382fd79.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 902.414837] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 902.416218] env[62814]: DEBUG nova.network.neutron [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Successfully created port: d5a713c4-180b-4ecd-9c1d-e2ad28f526ad {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 902.419146] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f57fcfe-588d-4ba1-8e05-5fe4915525c1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.429984] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 902.430260] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 902.431134] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff405847-afda-4f77-850d-e84cd4c1e179 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.438543] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 902.438543] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f52218-9d29-8f36-3d9e-617c2d9a82be" [ 902.438543] env[62814]: _type = "Task" [ 902.438543] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.448033] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f52218-9d29-8f36-3d9e-617c2d9a82be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.859975] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b51784-c29d-4839-ab3b-c40b5d14f157 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.869388] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Doing hard reboot of VM {{(pid=62814) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 902.869657] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-df9b3c24-43c9-4e56-97f4-2bd1404545f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.880696] env[62814]: DEBUG oslo_vmware.api [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 902.880696] env[62814]: value = "task-4293899" [ 902.880696] env[62814]: _type = "Task" [ 902.880696] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.891176] env[62814]: DEBUG oslo_vmware.api [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293899, 'name': ResetVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.927293] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 902.928088] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00152927-0728-4021-929a-ec23c36dad1f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.934684] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for the task: (returnval){ [ 902.934684] env[62814]: value = "task-4293900" [ 902.934684] env[62814]: _type = "Task" [ 902.934684] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.945960] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293900, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.953038] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Preparing fetch location {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 902.953038] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Fetch image to [datastore2] OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce/OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce.vmdk {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 902.953279] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Downloading stream optimized image 577d3b83-11e9-430b-a3da-3fc2f382fd79 to [datastore2] OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce/OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce.vmdk on the data store datastore2 as vApp {{(pid=62814) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 902.953526] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Downloading image file data 577d3b83-11e9-430b-a3da-3fc2f382fd79 to the ESX as VM named 'OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce' {{(pid=62814) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 903.012954] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.013298] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8eaf727d-acb1-4278-aae2-ad649d51ce6e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.021131] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 903.021131] env[62814]: value = "task-4293901" [ 903.021131] env[62814]: _type = "Task" [ 903.021131] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.034047] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293901, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.056263] env[62814]: DEBUG oslo_vmware.rw_handles [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 903.056263] env[62814]: value = "resgroup-9" [ 903.056263] env[62814]: _type = "ResourcePool" [ 903.056263] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 903.056621] env[62814]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-ff73d0f7-d362-406f-9e95-bbd2fe136cb4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.077843] env[62814]: DEBUG oslo_vmware.rw_handles [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lease: (returnval){ [ 903.077843] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272f0b7-190e-639c-660f-16ab67df7382" [ 903.077843] env[62814]: _type = "HttpNfcLease" [ 903.077843] env[62814]: } obtained for vApp import into resource pool (val){ [ 903.077843] env[62814]: value = "resgroup-9" [ 903.077843] env[62814]: _type = "ResourcePool" [ 903.077843] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 903.078258] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the lease: (returnval){ [ 903.078258] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272f0b7-190e-639c-660f-16ab67df7382" [ 903.078258] env[62814]: _type = "HttpNfcLease" [ 903.078258] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 903.084716] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 903.084716] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272f0b7-190e-639c-660f-16ab67df7382" [ 903.084716] env[62814]: _type = "HttpNfcLease" [ 903.084716] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 903.235979] env[62814]: DEBUG nova.compute.manager [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 903.239115] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Applying migration context for instance 9cd89da4-d190-4bfc-81e1-da15b98d54ba as it has an incoming, in-progress migration 8d6e5dee-326f-4a97-a651-a24c301bccc8. Migration status is reverting {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 903.242854] env[62814]: INFO nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating resource usage from migration 8d6e5dee-326f-4a97-a651-a24c301bccc8 [ 903.269802] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 903.269802] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 903.270437] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 903.270437] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 903.270437] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 903.270735] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 903.270782] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 903.270954] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 903.271820] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 903.272176] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 903.272384] env[62814]: DEBUG nova.virt.hardware [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 903.273293] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c218062d-1929-425f-b944-ac53c9d6c0a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.281438] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance c7cef7f3-11db-44e1-a454-98830b465b52 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.281438] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d9242042-6209-4b04-bf00-00dd04d9d6a0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.281576] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 6a592192-1b41-4be2-84a6-c3b76a4e5643 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.281694] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 59d21ef7-df97-49ac-9329-4c18df6dd087 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.281766] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 4ed66b36-b6c6-4673-9c03-169a01134574 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.281883] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance f5ad9d70-75fb-4881-8853-5ede4d0903f2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.282072] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance bd0933ca-aab7-4dd4-a570-1a58a720f377 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.282233] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 9343301c-dfe9-41b0-b4a0-067af544d297 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.282349] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d366a755-49b4-427b-8564-d8572a7fbbb7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.282459] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 3e4479a0-8edd-4b37-8cc9-2c91275b88ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.282570] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance cbc5985a-38e7-4e52-9fb0-264b5cec013a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.282683] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 1f40dc62-1a58-4cfb-8785-c37b68747f37 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.282793] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 425e8edd-c002-45a1-bb6f-ee3ac8812509 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.282906] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance c8765756-0870-4a06-a1a5-d02177959b29 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.283026] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 9683894b-a300-4400-a1b9-db62478f42c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.283171] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 903.283292] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance a473cc7e-0f86-4a18-9789-0aabfff430ab is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 903.283404] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance cd89e4d5-c08e-42c9-aea8-84c20714e196 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.283514] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 9be62576-5a05-473f-befd-b33f5fde9185 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 2, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.283627] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Migration 8d6e5dee-326f-4a97-a651-a24c301bccc8 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 903.283896] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 9cd89da4-d190-4bfc-81e1-da15b98d54ba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.284114] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 844737d8-d852-44bb-bf9d-e673e737ef33 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.284210] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 1e259ec6-d31a-453b-87e0-baa446665d56 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.284329] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 98884571-8101-4ae9-b2e3-aeed97e3618f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.284462] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 310e22c3-ff0a-4e6b-aa46-8812fa014dfa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.284596] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 98ba2786-023f-4d36-bdd0-e38bdd23d73c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.284724] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d7332882-19b3-4ab3-8ea6-51d33c584844 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.284860] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance af6ef867-dc9a-4db6-8582-a6198e5caa77 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.284974] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 7350d352-9336-40b8-81a6-0a4795d9f8dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 903.287446] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d244a9a-d5c8-4d16-a4a6-48f0d526c983 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.391929] env[62814]: DEBUG oslo_vmware.api [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293899, 'name': ResetVM_Task, 'duration_secs': 0.144108} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.392393] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Did hard reboot of VM {{(pid=62814) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 903.392734] env[62814]: DEBUG nova.compute.manager [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 903.393899] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdd387d-5a6b-4cb1-8180-536f0b39b8b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.444859] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293900, 'name': PowerOffVM_Task, 'duration_secs': 0.172775} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.445164] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 903.445405] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.446177] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f76ea83-7760-4d11-8aea-dea0accfddfe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.452902] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.453172] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7942c8d4-1006-4003-8f4a-7a8990c62cfa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.478428] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.478656] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.478860] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Deleting the datastore file [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.479320] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-338bd53b-b60e-4a4f-8c0c-82e84fe464b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.485873] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for the task: (returnval){ [ 903.485873] env[62814]: value = "task-4293904" [ 903.485873] env[62814]: _type = "Task" [ 903.485873] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.494659] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293904, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.531695] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293901, 'name': PowerOffVM_Task, 'duration_secs': 0.164457} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.531821] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 903.532496] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.533283] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ab0bed-ab6b-4169-abf2-de5f723f9075 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.540341] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.541065] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-599c69f5-c5b1-4630-9cec-d202849195fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.568713] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.568950] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.569211] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Deleting the datastore file [datastore2] d7332882-19b3-4ab3-8ea6-51d33c584844 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.569488] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7183f163-a5e7-4440-a421-04ff4de53cba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.576561] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 903.576561] env[62814]: value = "task-4293906" [ 903.576561] env[62814]: _type = "Task" [ 903.576561] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.586537] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 903.586537] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272f0b7-190e-639c-660f-16ab67df7382" [ 903.586537] env[62814]: _type = "HttpNfcLease" [ 903.586537] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 903.589824] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293906, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.793096] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 99864c68-5f11-4119-b8e3-3aa8719f267b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 903.912326] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ae1dec0c-5bf2-4641-b75c-dca679dbd2d3 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 5.047s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 903.998618] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293904, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.117739} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.998618] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.998618] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 903.998618] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.084140] env[62814]: DEBUG nova.compute.manager [req-afb617aa-a940-4f8c-9dfa-3d24822b81d2 req-532a7721-4141-48c1-beaa-7d42157e9926 service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Received event network-vif-plugged-d5a713c4-180b-4ecd-9c1d-e2ad28f526ad {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 904.084361] env[62814]: DEBUG oslo_concurrency.lockutils [req-afb617aa-a940-4f8c-9dfa-3d24822b81d2 req-532a7721-4141-48c1-beaa-7d42157e9926 service nova] Acquiring lock "7350d352-9336-40b8-81a6-0a4795d9f8dd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.084569] env[62814]: DEBUG oslo_concurrency.lockutils [req-afb617aa-a940-4f8c-9dfa-3d24822b81d2 req-532a7721-4141-48c1-beaa-7d42157e9926 service nova] Lock "7350d352-9336-40b8-81a6-0a4795d9f8dd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 904.084738] env[62814]: DEBUG oslo_concurrency.lockutils [req-afb617aa-a940-4f8c-9dfa-3d24822b81d2 req-532a7721-4141-48c1-beaa-7d42157e9926 service nova] Lock "7350d352-9336-40b8-81a6-0a4795d9f8dd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 904.084902] env[62814]: DEBUG nova.compute.manager [req-afb617aa-a940-4f8c-9dfa-3d24822b81d2 req-532a7721-4141-48c1-beaa-7d42157e9926 service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] No waiting events found dispatching network-vif-plugged-d5a713c4-180b-4ecd-9c1d-e2ad28f526ad {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 904.085286] env[62814]: WARNING nova.compute.manager [req-afb617aa-a940-4f8c-9dfa-3d24822b81d2 req-532a7721-4141-48c1-beaa-7d42157e9926 service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Received unexpected event network-vif-plugged-d5a713c4-180b-4ecd-9c1d-e2ad28f526ad for instance with vm_state building and task_state spawning. [ 904.093894] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293906, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113267} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.095430] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.095641] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.095789] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.098101] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 904.098101] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272f0b7-190e-639c-660f-16ab67df7382" [ 904.098101] env[62814]: _type = "HttpNfcLease" [ 904.098101] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 904.098944] env[62814]: DEBUG oslo_vmware.rw_handles [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 904.098944] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272f0b7-190e-639c-660f-16ab67df7382" [ 904.098944] env[62814]: _type = "HttpNfcLease" [ 904.098944] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 904.099665] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea02cfe-a328-4466-87a7-3daf31cd3823 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.107664] env[62814]: DEBUG oslo_vmware.rw_handles [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d2a59-dca2-e277-cf73-6ce6338e4b52/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 904.107846] env[62814]: DEBUG oslo_vmware.rw_handles [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d2a59-dca2-e277-cf73-6ce6338e4b52/disk-0.vmdk. {{(pid=62814) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 904.168131] env[62814]: DEBUG nova.network.neutron [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Successfully updated port: d5a713c4-180b-4ecd-9c1d-e2ad28f526ad {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.174195] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6a8981a6-0a47-4dfb-bf54-14ebb5c382ad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.296818] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 904.670940] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "refresh_cache-7350d352-9336-40b8-81a6-0a4795d9f8dd" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.670940] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "refresh_cache-7350d352-9336-40b8-81a6-0a4795d9f8dd" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 904.670940] env[62814]: DEBUG nova.network.neutron [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.799880] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 5c574786-64a4-4f07-a267-101ecaaa6938 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 904.800246] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 27 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 904.800419] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=5824MB phys_disk=149GB used_disk=27GB total_vcpus=48 used_vcpus=27 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 905.053211] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 905.053211] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.053211] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 905.053627] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.053627] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 905.053807] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 905.053895] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 905.054155] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 905.054328] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 905.054581] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 905.055042] env[62814]: DEBUG nova.virt.hardware [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 905.055931] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f70011e-5673-4594-a0e1-c918ebbb88d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.067044] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e353210d-2a97-4e57-905c-abd29a564140 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.088608] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.094695] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 905.104250] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.110306] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-815794d1-8bf0-490b-9d3f-9cf95a6843b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.137863] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.137863] env[62814]: value = "task-4293907" [ 905.137863] env[62814]: _type = "Task" [ 905.137863] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.145109] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 905.145364] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 905.145510] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 905.145686] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 905.145826] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 905.145968] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 905.146194] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 905.146415] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 905.147029] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 905.147029] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 905.147029] env[62814]: DEBUG nova.virt.hardware [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 905.153716] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec9d51f-18db-4838-9d77-f1f261578658 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.161956] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293907, 'name': CreateVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.170326] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d24737c-c6da-4a40-903c-580f5af45797 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.191199] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.197065] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 905.204270] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.204703] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3ff5ab69-3f74-4433-a51c-5f7dafd8c392 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.228250] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.228250] env[62814]: value = "task-4293908" [ 905.228250] env[62814]: _type = "Task" [ 905.228250] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.241248] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293908, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.244277] env[62814]: DEBUG nova.network.neutron [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 905.284109] env[62814]: DEBUG nova.compute.manager [req-15090f74-521c-4a0c-a7af-6aad44690514 req-2fe3de7d-9238-47d2-909a-cc438cd44fd8 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Received event network-changed-5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 905.284463] env[62814]: DEBUG nova.compute.manager [req-15090f74-521c-4a0c-a7af-6aad44690514 req-2fe3de7d-9238-47d2-909a-cc438cd44fd8 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Refreshing instance network info cache due to event network-changed-5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 905.284559] env[62814]: DEBUG oslo_concurrency.lockutils [req-15090f74-521c-4a0c-a7af-6aad44690514 req-2fe3de7d-9238-47d2-909a-cc438cd44fd8 service nova] Acquiring lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.284720] env[62814]: DEBUG oslo_concurrency.lockutils [req-15090f74-521c-4a0c-a7af-6aad44690514 req-2fe3de7d-9238-47d2-909a-cc438cd44fd8 service nova] Acquired lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 905.284942] env[62814]: DEBUG nova.network.neutron [req-15090f74-521c-4a0c-a7af-6aad44690514 req-2fe3de7d-9238-47d2-909a-cc438cd44fd8 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Refreshing network info cache for port 5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 905.314342] env[62814]: DEBUG oslo_vmware.rw_handles [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Completed reading data from the image iterator. {{(pid=62814) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 905.314576] env[62814]: DEBUG oslo_vmware.rw_handles [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d2a59-dca2-e277-cf73-6ce6338e4b52/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 905.315944] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9183b13a-1ab2-431f-9522-786f3e7b155b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.324511] env[62814]: DEBUG oslo_vmware.rw_handles [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d2a59-dca2-e277-cf73-6ce6338e4b52/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 905.324511] env[62814]: DEBUG oslo_vmware.rw_handles [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d2a59-dca2-e277-cf73-6ce6338e4b52/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 905.326741] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-897af9bf-3cf7-4734-8268-c0872b9524a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.500024] env[62814]: DEBUG nova.network.neutron [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Updating instance_info_cache with network_info: [{"id": "d5a713c4-180b-4ecd-9c1d-e2ad28f526ad", "address": "fa:16:3e:76:f9:fc", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a713c4-18", "ovs_interfaceid": "d5a713c4-180b-4ecd-9c1d-e2ad28f526ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.516552] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c91b72-fe67-4f94-bd08-4ae571b9d9b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.526219] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520c43ef-4bb1-4d52-810d-f36d80517c9d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.565593] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d9c2e5-75d9-4747-95a3-d7ecf9b6e42b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.568801] env[62814]: DEBUG oslo_vmware.rw_handles [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522d2a59-dca2-e277-cf73-6ce6338e4b52/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 905.569150] env[62814]: INFO nova.virt.vmwareapi.images [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Downloaded image file data 577d3b83-11e9-430b-a3da-3fc2f382fd79 [ 905.569893] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a46ab15-3ba4-4506-aff7-18fabf49a7db {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.589521] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0997ef56-9da5-4ed5-b650-c7c2f2ce3e25 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.591974] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36c64560-87f6-4e41-90d9-38d1e45c341e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.608349] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 905.652353] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293907, 'name': CreateVM_Task, 'duration_secs': 0.384439} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.652514] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 905.652922] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.653082] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 905.653400] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 905.653657] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbb67494-dcb2-40ef-b13b-32d11d38bd83 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.658046] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for the task: (returnval){ [ 905.658046] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5210949a-962c-ffbe-a2c6-2d81e56eb839" [ 905.658046] env[62814]: _type = "Task" [ 905.658046] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.666030] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5210949a-962c-ffbe-a2c6-2d81e56eb839, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.702190] env[62814]: INFO nova.virt.vmwareapi.images [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] The imported VM was unregistered [ 905.705036] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Caching image {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 905.705303] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating directory with path [datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79 {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 905.705583] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-93fab7f4-41ae-4068-8368-401448f318d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.720816] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Created directory with path [datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79 {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 905.721040] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce/OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce.vmdk to [datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79/577d3b83-11e9-430b-a3da-3fc2f382fd79.vmdk. {{(pid=62814) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 905.722047] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-300d1e78-10ab-4cb1-a310-8b45caded305 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.728601] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 905.728601] env[62814]: value = "task-4293910" [ 905.728601] env[62814]: _type = "Task" [ 905.728601] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.741377] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293908, 'name': CreateVM_Task, 'duration_secs': 0.316004} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.745252] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 905.745541] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293910, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.745900] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.949159] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 905.949552] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.949846] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 905.950112] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.950354] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.953368] env[62814]: INFO nova.compute.manager [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Terminating instance [ 906.003277] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "refresh_cache-7350d352-9336-40b8-81a6-0a4795d9f8dd" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.003743] env[62814]: DEBUG nova.compute.manager [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Instance network_info: |[{"id": "d5a713c4-180b-4ecd-9c1d-e2ad28f526ad", "address": "fa:16:3e:76:f9:fc", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a713c4-18", "ovs_interfaceid": "d5a713c4-180b-4ecd-9c1d-e2ad28f526ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 906.004092] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:76:f9:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5a713c4-180b-4ecd-9c1d-e2ad28f526ad', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.011978] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Creating folder: Project (d82993ef7dfa4d2f8f39db3577dd321a). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 906.015115] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e536cd0f-9adf-47bf-8d33-98d92205daee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.026698] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Created folder: Project (d82993ef7dfa4d2f8f39db3577dd321a) in parent group-v845547. [ 906.026698] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Creating folder: Instances. Parent ref: group-v845730. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 906.026995] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-80d218fa-587e-45e4-af27-98c229e53e59 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.036756] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Created folder: Instances in parent group-v845730. [ 906.037021] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 906.037217] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 906.037426] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5da0260-7d22-423b-aab8-67c4722803ee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.058758] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.058758] env[62814]: value = "task-4293913" [ 906.058758] env[62814]: _type = "Task" [ 906.058758] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.068895] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293913, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.116083] env[62814]: DEBUG nova.network.neutron [req-15090f74-521c-4a0c-a7af-6aad44690514 req-2fe3de7d-9238-47d2-909a-cc438cd44fd8 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Updated VIF entry in instance network info cache for port 5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.116453] env[62814]: DEBUG nova.network.neutron [req-15090f74-521c-4a0c-a7af-6aad44690514 req-2fe3de7d-9238-47d2-909a-cc438cd44fd8 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Updating instance_info_cache with network_info: [{"id": "5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca", "address": "fa:16:3e:aa:06:a1", "network": {"id": "4ddbe620-bc9d-4b38-a414-7799c4e7060d", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-334880125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ef503fd02adc475fa79dda8b379b6514", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c964bdc6-fccc-40d9-bfe2-763b6f05a863", "external-id": "cl2-zone-376", "segmentation_id": 376, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e2b0ae5-35", "ovs_interfaceid": "5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.120257] env[62814]: DEBUG nova.compute.manager [req-2418a390-ba5d-4cbb-bb6a-3634c27d14b4 req-638f55b8-4a0e-44c2-8286-ddbc86e6e58c service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Received event network-changed-d5a713c4-180b-4ecd-9c1d-e2ad28f526ad {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 906.120574] env[62814]: DEBUG nova.compute.manager [req-2418a390-ba5d-4cbb-bb6a-3634c27d14b4 req-638f55b8-4a0e-44c2-8286-ddbc86e6e58c service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Refreshing instance network info cache due to event network-changed-d5a713c4-180b-4ecd-9c1d-e2ad28f526ad. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 906.120901] env[62814]: DEBUG oslo_concurrency.lockutils [req-2418a390-ba5d-4cbb-bb6a-3634c27d14b4 req-638f55b8-4a0e-44c2-8286-ddbc86e6e58c service nova] Acquiring lock "refresh_cache-7350d352-9336-40b8-81a6-0a4795d9f8dd" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.121188] env[62814]: DEBUG oslo_concurrency.lockutils [req-2418a390-ba5d-4cbb-bb6a-3634c27d14b4 req-638f55b8-4a0e-44c2-8286-ddbc86e6e58c service nova] Acquired lock "refresh_cache-7350d352-9336-40b8-81a6-0a4795d9f8dd" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.121499] env[62814]: DEBUG nova.network.neutron [req-2418a390-ba5d-4cbb-bb6a-3634c27d14b4 req-638f55b8-4a0e-44c2-8286-ddbc86e6e58c service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Refreshing network info cache for port d5a713c4-180b-4ecd-9c1d-e2ad28f526ad {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 906.133954] env[62814]: ERROR nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [req-d1558923-efb2-46d5-be6c-bed527e80e4b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d1558923-efb2-46d5-be6c-bed527e80e4b"}]} [ 906.153391] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 906.170363] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5210949a-962c-ffbe-a2c6-2d81e56eb839, 'name': SearchDatastore_Task, 'duration_secs': 0.010108} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.170463] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.170883] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.171276] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.171505] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.171778] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.173613] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 906.173912] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 906.176171] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.176502] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 906.176766] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bb56cb4a-eda6-4a77-9463-eee4b5c7804c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.179448] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aae39b09-ca8e-4be8-acec-932ffe6e8e57 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.189812] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 906.193889] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 906.193889] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e92810-2e90-6746-ade7-792c61001ff4" [ 906.193889] env[62814]: _type = "Task" [ 906.193889] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.204084] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.204253] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.205514] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-226f9b80-5692-4de6-ac93-78e3a2506d28 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.212520] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e92810-2e90-6746-ade7-792c61001ff4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.214862] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for the task: (returnval){ [ 906.214862] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52451203-57d7-91b5-082e-881a9fba65e7" [ 906.214862] env[62814]: _type = "Task" [ 906.214862] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.224223] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52451203-57d7-91b5-082e-881a9fba65e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.224710] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 906.238947] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293910, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.462967] env[62814]: DEBUG nova.compute.manager [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 906.462967] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.463283] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b937778e-50ab-4977-9954-f86db38fbfa0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.473332] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.473617] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-687824f1-4ae2-4a0c-a5e0-f6a60662bba6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.483852] env[62814]: DEBUG oslo_vmware.api [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 906.483852] env[62814]: value = "task-4293914" [ 906.483852] env[62814]: _type = "Task" [ 906.483852] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.492727] env[62814]: DEBUG oslo_vmware.api [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293914, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.572586] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293913, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.624185] env[62814]: DEBUG oslo_concurrency.lockutils [req-15090f74-521c-4a0c-a7af-6aad44690514 req-2fe3de7d-9238-47d2-909a-cc438cd44fd8 service nova] Releasing lock "refresh_cache-310e22c3-ff0a-4e6b-aa46-8812fa014dfa" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.688254] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquiring lock "1f40dc62-1a58-4cfb-8785-c37b68747f37" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.688254] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lock "1f40dc62-1a58-4cfb-8785-c37b68747f37" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.688254] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquiring lock "1f40dc62-1a58-4cfb-8785-c37b68747f37-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 906.688254] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lock "1f40dc62-1a58-4cfb-8785-c37b68747f37-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.688254] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lock "1f40dc62-1a58-4cfb-8785-c37b68747f37-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.691249] env[62814]: INFO nova.compute.manager [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Terminating instance [ 906.708252] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e92810-2e90-6746-ade7-792c61001ff4, 'name': SearchDatastore_Task, 'duration_secs': 0.090657} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.709540] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.709816] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.710104] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.712751] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aeae79e-2e60-4a0f-9537-9362d25bba1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.729870] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56551358-3ac3-4325-ae11-c6d06ae1a127 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.734531] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52451203-57d7-91b5-082e-881a9fba65e7, 'name': SearchDatastore_Task, 'duration_secs': 0.092329} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.739339] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dee1067-918d-4b6a-8d2d-6cbe9972b9f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.782607] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8488cd-eae0-453c-add0-b5f7e57c95d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.785604] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for the task: (returnval){ [ 906.785604] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527db2ad-bc3b-1410-ea40-29e2169783e3" [ 906.785604] env[62814]: _type = "Task" [ 906.785604] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.785891] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293910, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.794053] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a846ff-7afe-4914-b63b-18a8083272f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.802151] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527db2ad-bc3b-1410-ea40-29e2169783e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.814766] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 906.927900] env[62814]: DEBUG nova.network.neutron [req-2418a390-ba5d-4cbb-bb6a-3634c27d14b4 req-638f55b8-4a0e-44c2-8286-ddbc86e6e58c service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Updated VIF entry in instance network info cache for port d5a713c4-180b-4ecd-9c1d-e2ad28f526ad. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.928893] env[62814]: DEBUG nova.network.neutron [req-2418a390-ba5d-4cbb-bb6a-3634c27d14b4 req-638f55b8-4a0e-44c2-8286-ddbc86e6e58c service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Updating instance_info_cache with network_info: [{"id": "d5a713c4-180b-4ecd-9c1d-e2ad28f526ad", "address": "fa:16:3e:76:f9:fc", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5a713c4-18", "ovs_interfaceid": "d5a713c4-180b-4ecd-9c1d-e2ad28f526ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.994380] env[62814]: DEBUG oslo_vmware.api [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293914, 'name': PowerOffVM_Task, 'duration_secs': 0.467005} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.994706] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.994932] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.995374] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef9e2284-c07e-4643-9be2-19604b9d9df8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.074017] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293913, 'name': CreateVM_Task, 'duration_secs': 0.744862} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.076681] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 907.080023] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 907.080023] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 907.080023] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Deleting the datastore file [datastore2] 310e22c3-ff0a-4e6b-aa46-8812fa014dfa {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 907.080023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.080023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 907.080023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 907.080023] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-83a8900c-c586-459c-857a-dbd85d7e6ce2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.081726] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a60be22e-89d8-42a1-aeaf-a23418f31942 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.089326] env[62814]: DEBUG oslo_vmware.api [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 907.089326] env[62814]: value = "task-4293916" [ 907.089326] env[62814]: _type = "Task" [ 907.089326] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.090451] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 907.090451] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5280cf61-449a-30a3-1ff4-c6159200b8de" [ 907.090451] env[62814]: _type = "Task" [ 907.090451] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.101503] env[62814]: DEBUG oslo_vmware.api [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293916, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.104776] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5280cf61-449a-30a3-1ff4-c6159200b8de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.203437] env[62814]: DEBUG nova.compute.manager [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 907.203437] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 907.204369] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a661dc2-7889-4ae4-ac50-ed14a70daeff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.212833] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 907.213020] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8491e6cc-cd43-42e5-bd40-289080bbac74 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.220130] env[62814]: DEBUG oslo_vmware.api [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for the task: (returnval){ [ 907.220130] env[62814]: value = "task-4293917" [ 907.220130] env[62814]: _type = "Task" [ 907.220130] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.229957] env[62814]: DEBUG oslo_vmware.api [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.244342] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293910, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.296503] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527db2ad-bc3b-1410-ea40-29e2169783e3, 'name': SearchDatastore_Task, 'duration_secs': 0.082734} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.296666] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.296961] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.297294] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 907.297486] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.297719] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bbbee21b-14e7-41b4-958a-e84b3171e119 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.300857] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fafd64eb-2728-47d1-9684-1f4a34cb3a78 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.307687] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for the task: (returnval){ [ 907.307687] env[62814]: value = "task-4293918" [ 907.307687] env[62814]: _type = "Task" [ 907.307687] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.317425] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293918, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.321220] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.321411] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 907.322401] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aee96111-1372-4731-9165-b1ab3cbb0b0b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.328360] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 907.328360] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d961de-cb8e-bd0b-e2cb-7af75f9db7c4" [ 907.328360] env[62814]: _type = "Task" [ 907.328360] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.336938] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d961de-cb8e-bd0b-e2cb-7af75f9db7c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.359900] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 91 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 907.360214] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 91 to 92 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 907.360378] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 907.432088] env[62814]: DEBUG oslo_concurrency.lockutils [req-2418a390-ba5d-4cbb-bb6a-3634c27d14b4 req-638f55b8-4a0e-44c2-8286-ddbc86e6e58c service nova] Releasing lock "refresh_cache-7350d352-9336-40b8-81a6-0a4795d9f8dd" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.604275] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5280cf61-449a-30a3-1ff4-c6159200b8de, 'name': SearchDatastore_Task, 'duration_secs': 0.081091} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.607575] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.607835] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.608113] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.608369] env[62814]: DEBUG oslo_vmware.api [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293916, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.731507] env[62814]: DEBUG oslo_vmware.api [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.750022] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293910, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.818277] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293918, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.839871] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d961de-cb8e-bd0b-e2cb-7af75f9db7c4, 'name': SearchDatastore_Task, 'duration_secs': 0.090663} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.840850] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86aeaea2-d8fa-4b26-bf92-dca98a2cd6a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.849942] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 907.849942] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52da4a85-bdac-46df-fe7e-cab5481ba208" [ 907.849942] env[62814]: _type = "Task" [ 907.849942] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.858784] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52da4a85-bdac-46df-fe7e-cab5481ba208, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.865717] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 907.866833] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.650s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 907.866833] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.330s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 907.866833] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 907.868633] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.805s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 907.870335] env[62814]: INFO nova.compute.claims [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 907.873543] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 907.873722] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Cleaning up deleted instances {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11865}} [ 907.905464] env[62814]: INFO nova.scheduler.client.report [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Deleted allocations for instance a473cc7e-0f86-4a18-9789-0aabfff430ab [ 908.101905] env[62814]: DEBUG oslo_vmware.api [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293916, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.232657] env[62814]: DEBUG oslo_vmware.api [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293917, 'name': PowerOffVM_Task, 'duration_secs': 0.97105} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.232943] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 908.233129] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 908.233395] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7d510389-2196-4e4e-bf2a-f84217391f87 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.246691] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293910, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.499657} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.246691] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce/OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce.vmdk to [datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79/577d3b83-11e9-430b-a3da-3fc2f382fd79.vmdk. [ 908.246842] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Cleaning up location [datastore2] OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 908.246989] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_0b3b7fab-6e0a-46f5-88d1-03ee8308a2ce {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 908.247470] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b0647b4-61d2-4f6f-ab81-c114c42c8300 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.252915] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 908.252915] env[62814]: value = "task-4293920" [ 908.252915] env[62814]: _type = "Task" [ 908.252915] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.260798] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293920, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.317967] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293918, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.319199] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 908.319404] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 908.319581] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Deleting the datastore file [datastore2] 1f40dc62-1a58-4cfb-8785-c37b68747f37 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 908.319834] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1439cc0c-453b-4aa3-8b95-0bed4f360c97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.325337] env[62814]: DEBUG oslo_vmware.api [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for the task: (returnval){ [ 908.325337] env[62814]: value = "task-4293921" [ 908.325337] env[62814]: _type = "Task" [ 908.325337] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.335429] env[62814]: DEBUG oslo_vmware.api [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.361746] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52da4a85-bdac-46df-fe7e-cab5481ba208, 'name': SearchDatastore_Task, 'duration_secs': 0.086824} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.362089] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 908.362420] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d7332882-19b3-4ab3-8ea6-51d33c584844/d7332882-19b3-4ab3-8ea6-51d33c584844.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 908.362682] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 908.362936] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 908.363204] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-57f62caf-b3c5-430a-a14e-df6cbf817951 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.365525] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b06a161-c003-4dab-81b8-e48de1c9619f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.372057] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 908.372057] env[62814]: value = "task-4293922" [ 908.372057] env[62814]: _type = "Task" [ 908.372057] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.388489] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] There are 38 instances to clean {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11874}} [ 908.388643] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 448b7bfb-1a90-4c5b-9086-9320cedaf015] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 908.393129] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 908.393373] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 908.398752] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82aaf19f-6a03-462a-b739-4bdbed5d6bf7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.400447] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.404257] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 908.404257] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525fe5ff-1210-6b3a-d3f7-7538c8f0660a" [ 908.404257] env[62814]: _type = "Task" [ 908.404257] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.418041] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525fe5ff-1210-6b3a-d3f7-7538c8f0660a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.418041] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7ccd5ff8-2ea8-4066-8e1e-b8ddc9dbcf11 tempest-ServersTestMultiNic-1834266017 tempest-ServersTestMultiNic-1834266017-project-member] Lock "a473cc7e-0f86-4a18-9789-0aabfff430ab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.551s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.603497] env[62814]: DEBUG oslo_vmware.api [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4293916, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.154639} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.604043] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 908.605688] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 908.606025] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 908.608035] env[62814]: INFO nova.compute.manager [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Took 2.14 seconds to destroy the instance on the hypervisor. [ 908.608035] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 908.608035] env[62814]: DEBUG nova.compute.manager [-] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 908.608035] env[62814]: DEBUG nova.network.neutron [-] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 908.764522] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293920, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10653} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.764917] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 908.765215] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "[datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79/577d3b83-11e9-430b-a3da-3fc2f382fd79.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 908.765608] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79/577d3b83-11e9-430b-a3da-3fc2f382fd79.vmdk to [datastore2] af6ef867-dc9a-4db6-8582-a6198e5caa77/af6ef867-dc9a-4db6-8582-a6198e5caa77.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 908.766365] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8396f1b7-3b3c-408c-943a-ed191e2de475 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.775579] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 908.775579] env[62814]: value = "task-4293923" [ 908.775579] env[62814]: _type = "Task" [ 908.775579] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.784559] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293923, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.820850] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293918, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.434526} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.820850] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.821142] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.821442] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a44328ea-fc84-45e2-96ce-e510fc689fbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.835279] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for the task: (returnval){ [ 908.835279] env[62814]: value = "task-4293924" [ 908.835279] env[62814]: _type = "Task" [ 908.835279] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.842879] env[62814]: DEBUG oslo_vmware.api [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.848612] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293924, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.882425] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.900069] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 413eaa2e-7bbc-402e-b0d3-f030b6395d7b] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 908.929538] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525fe5ff-1210-6b3a-d3f7-7538c8f0660a, 'name': SearchDatastore_Task, 'duration_secs': 0.058548} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.929538] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30446858-1ba7-4bda-a2b8-e44041786e02 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.934377] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 908.934377] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524c922f-319c-9ea3-1c8d-f6fe279ebeaa" [ 908.934377] env[62814]: _type = "Task" [ 908.934377] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.948961] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524c922f-319c-9ea3-1c8d-f6fe279ebeaa, 'name': SearchDatastore_Task, 'duration_secs': 0.010159} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.949309] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 908.949519] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 7350d352-9336-40b8-81a6-0a4795d9f8dd/7350d352-9336-40b8-81a6-0a4795d9f8dd.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 908.949842] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-546d4771-4209-40b9-9cdf-5df5eb650762 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.959879] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 908.959879] env[62814]: value = "task-4293925" [ 908.959879] env[62814]: _type = "Task" [ 908.959879] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.972887] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293925, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.067275] env[62814]: DEBUG nova.compute.manager [req-2cecae85-5336-43e5-b7e0-75c4139032ac req-57b9a848-f06a-4b5c-867c-5548afd2bb25 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Received event network-vif-deleted-5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 909.067460] env[62814]: INFO nova.compute.manager [req-2cecae85-5336-43e5-b7e0-75c4139032ac req-57b9a848-f06a-4b5c-867c-5548afd2bb25 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Neutron deleted interface 5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca; detaching it from the instance and deleting it from the info cache [ 909.067627] env[62814]: DEBUG nova.network.neutron [req-2cecae85-5336-43e5-b7e0-75c4139032ac req-57b9a848-f06a-4b5c-867c-5548afd2bb25 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.286929] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293923, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.343876] env[62814]: DEBUG oslo_vmware.api [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Task: {'id': task-4293921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.653057} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.351100] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 909.351458] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 909.351868] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 909.352151] env[62814]: INFO nova.compute.manager [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Took 2.15 seconds to destroy the instance on the hypervisor. [ 909.352486] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 909.353584] env[62814]: DEBUG nova.compute.manager [-] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 909.353744] env[62814]: DEBUG nova.network.neutron [-] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 909.362195] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293924, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075531} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.362497] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.363333] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a29b98e-d6c1-4465-a886-90a0037874af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.386158] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.390537] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08e6db54-930b-47ca-bd52-14d0e1c9bdda {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.409278] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 7cca1c35-6bfc-450d-ba74-0e825b160e8f] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 909.419024] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.421014] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for the task: (returnval){ [ 909.421014] env[62814]: value = "task-4293926" [ 909.421014] env[62814]: _type = "Task" [ 909.421014] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.429918] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293926, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.463289] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8ebae5-053d-416b-aff6-5ef0b6d57ab8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.477785] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-090c9cef-e21e-4b6e-8076-f65e63ecd61b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.481061] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293925, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.508946] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14963961-606e-4b66-a4af-c16fd658e39b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.516268] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e8b7e4f-bf90-4bfb-be9e-835b416b9160 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.530219] env[62814]: DEBUG nova.compute.provider_tree [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 909.541163] env[62814]: DEBUG nova.network.neutron [-] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.570998] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9aa17998-50fd-4d89-a3d0-c4d3963ce6e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.583021] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d624e42f-b19e-4d53-920d-0235e4a6c3b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.621500] env[62814]: DEBUG nova.compute.manager [req-2cecae85-5336-43e5-b7e0-75c4139032ac req-57b9a848-f06a-4b5c-867c-5548afd2bb25 service nova] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Detach interface failed, port_id=5e2b0ae5-3528-47ac-b7c7-cc8cc800c1ca, reason: Instance 310e22c3-ff0a-4e6b-aa46-8812fa014dfa could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 909.787793] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293923, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.856988] env[62814]: DEBUG oslo_concurrency.lockutils [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquiring lock "cd89e4d5-c08e-42c9-aea8-84c20714e196" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 909.856988] env[62814]: DEBUG oslo_concurrency.lockutils [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lock "cd89e4d5-c08e-42c9-aea8-84c20714e196" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 909.857119] env[62814]: DEBUG oslo_concurrency.lockutils [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquiring lock "cd89e4d5-c08e-42c9-aea8-84c20714e196-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 909.857903] env[62814]: DEBUG oslo_concurrency.lockutils [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lock "cd89e4d5-c08e-42c9-aea8-84c20714e196-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 909.857903] env[62814]: DEBUG oslo_concurrency.lockutils [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lock "cd89e4d5-c08e-42c9-aea8-84c20714e196-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.863820] env[62814]: INFO nova.compute.manager [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Terminating instance [ 909.889822] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293922, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.915270] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: c18acec7-cf95-4cdf-aa49-32419d364534] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 909.936366] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.983829] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293925, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.042530] env[62814]: INFO nova.compute.manager [-] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Took 1.44 seconds to deallocate network for instance. [ 910.093120] env[62814]: DEBUG nova.scheduler.client.report [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 92 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 910.093444] env[62814]: DEBUG nova.compute.provider_tree [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 92 to 93 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 910.093635] env[62814]: DEBUG nova.compute.provider_tree [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 910.288484] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293923, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.327363] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.331088] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.370214] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Acquiring lock "bd0933ca-aab7-4dd4-a570-1a58a720f377" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.370214] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lock "bd0933ca-aab7-4dd4-a570-1a58a720f377" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.370214] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Acquiring lock "bd0933ca-aab7-4dd4-a570-1a58a720f377-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.370214] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lock "bd0933ca-aab7-4dd4-a570-1a58a720f377-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.370214] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lock "bd0933ca-aab7-4dd4-a570-1a58a720f377-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.371611] env[62814]: DEBUG nova.compute.manager [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 910.371839] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 910.372445] env[62814]: INFO nova.compute.manager [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Terminating instance [ 910.374750] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0972b534-0a52-47f1-ad5e-f6f45372cac1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.395020] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293922, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.395915] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.396405] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b49b1b4-bf5c-4bca-9575-49af6894296a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.404374] env[62814]: DEBUG oslo_vmware.api [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for the task: (returnval){ [ 910.404374] env[62814]: value = "task-4293927" [ 910.404374] env[62814]: _type = "Task" [ 910.404374] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.415560] env[62814]: DEBUG oslo_vmware.api [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.423039] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 2d0b24c5-3593-4ef2-a637-d3590242ad79] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 910.437209] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.479170] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293925, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.518152] env[62814]: DEBUG nova.network.neutron [-] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.550862] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 910.602136] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.733s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.602394] env[62814]: DEBUG nova.compute.manager [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 910.606526] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 29.320s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 910.607384] env[62814]: DEBUG nova.objects.instance [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 910.789785] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293923, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.835520] env[62814]: DEBUG nova.compute.utils [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 910.885652] env[62814]: DEBUG nova.compute.manager [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 910.885929] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 910.886356] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e24c5826-fd6e-431d-a5af-8f410f94e763 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.895084] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293922, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.896743] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for the task: (returnval){ [ 910.896743] env[62814]: value = "task-4293928" [ 910.896743] env[62814]: _type = "Task" [ 910.896743] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.909329] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293928, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.915582] env[62814]: DEBUG oslo_vmware.api [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.929735] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 127fee64-fd56-4a23-bdd2-18c817898fd5] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 910.940059] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.982494] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293925, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.022786] env[62814]: INFO nova.compute.manager [-] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Took 1.67 seconds to deallocate network for instance. [ 911.115224] env[62814]: DEBUG nova.compute.utils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 911.122029] env[62814]: DEBUG nova.compute.manager [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 911.122029] env[62814]: DEBUG nova.network.neutron [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 911.182538] env[62814]: DEBUG nova.policy [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1d52f0fcc0ec42d69e6b2094145d0f28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '558682c8ea0f4887874a1763f65cb9a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 911.235183] env[62814]: DEBUG nova.compute.manager [req-dcc6e30e-914e-4b22-bdfc-5c27dac87820 req-c2fa104a-0875-486c-b65b-0834dc8040f9 service nova] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Received event network-vif-deleted-f5994178-e05c-464f-afea-a4f8a606efe3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 911.290561] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293923, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.343078] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.015s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 911.397599] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293922, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.607385} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.397599] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d7332882-19b3-4ab3-8ea6-51d33c584844/d7332882-19b3-4ab3-8ea6-51d33c584844.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.397744] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.403165] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51af8666-aabe-48d8-b701-b7f67b1a662f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.417192] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293928, 'name': PowerOffVM_Task, 'duration_secs': 0.441496} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.417576] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 911.417576] env[62814]: value = "task-4293929" [ 911.417576] env[62814]: _type = "Task" [ 911.417576] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.418829] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 911.422280] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 911.422692] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845587', 'volume_id': 'fa377baf-042f-48ba-8dc4-21baec93c237', 'name': 'volume-fa377baf-042f-48ba-8dc4-21baec93c237', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bd0933ca-aab7-4dd4-a570-1a58a720f377', 'attached_at': '', 'detached_at': '', 'volume_id': 'fa377baf-042f-48ba-8dc4-21baec93c237', 'serial': 'fa377baf-042f-48ba-8dc4-21baec93c237'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 911.429474] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec4e599-0bee-495d-b383-e2dd59170cad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.437558] env[62814]: DEBUG oslo_vmware.api [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.443248] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: b289800f-49ce-49e7-b6bc-a3b4ec84b434] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 911.450466] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293929, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.470223] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb8c614-55a8-449a-bdfd-b2245df35ff5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.477953] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293926, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.481675] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5159cf-d900-4f9d-99d1-408827160c98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.488799] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293925, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.045837} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.490088] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 7350d352-9336-40b8-81a6-0a4795d9f8dd/7350d352-9336-40b8-81a6-0a4795d9f8dd.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.490400] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.491390] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e317c030-1f81-489c-8122-d885c8deeae3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.508684] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15832d35-b8be-43d8-b0df-fa4787f44c3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.515279] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 911.515279] env[62814]: value = "task-4293930" [ 911.515279] env[62814]: _type = "Task" [ 911.515279] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.528741] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] The volume has not been displaced from its original location: [datastore2] volume-fa377baf-042f-48ba-8dc4-21baec93c237/volume-fa377baf-042f-48ba-8dc4-21baec93c237.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 911.537190] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Reconfiguring VM instance instance-0000001e to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 911.539057] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 911.539332] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2856c83-d1ff-4c5f-a382-232fa7daecac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.558923] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293930, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.561125] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for the task: (returnval){ [ 911.561125] env[62814]: value = "task-4293931" [ 911.561125] env[62814]: _type = "Task" [ 911.561125] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.572154] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293931, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.620307] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b354ab78-75f4-4949-b228-c15c10969e16 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 911.621623] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.361s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 911.623904] env[62814]: INFO nova.compute.claims [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.626518] env[62814]: DEBUG nova.compute.manager [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 911.688155] env[62814]: DEBUG nova.network.neutron [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Successfully created port: 5764c0a4-20c2-4baa-a42a-0968ba00981e {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.792742] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293923, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.923755] env[62814]: DEBUG oslo_vmware.api [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293927, 'name': PowerOffVM_Task, 'duration_secs': 1.266586} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.924255] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 911.924452] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 911.929616] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8bb88484-1604-4790-97d5-9e6be9848796 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.940679] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293929, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.33022} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.944084] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.944987] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfd9510c-f8b6-42ed-8c04-5ac8af43b9f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.963022] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: c9e11016-e92a-459e-b5ee-b0e43ce29450] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 911.972610] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] d7332882-19b3-4ab3-8ea6-51d33c584844/d7332882-19b3-4ab3-8ea6-51d33c584844.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.977341] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80b912cf-9c4c-4a17-91aa-fd8fc91fa9b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.995168] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293926, 'name': ReconfigVM_Task, 'duration_secs': 2.487629} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.995501] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f/98884571-8101-4ae9-b2e3-aeed97e3618f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 911.996834] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1a8458f-c318-4e69-b59b-0ae5bb237481 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.002750] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 912.002750] env[62814]: value = "task-4293933" [ 912.002750] env[62814]: _type = "Task" [ 912.002750] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.007827] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for the task: (returnval){ [ 912.007827] env[62814]: value = "task-4293934" [ 912.007827] env[62814]: _type = "Task" [ 912.007827] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.019220] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293934, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.027515] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293933, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.043039] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293930, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.321857} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.043899] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 912.044299] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca74b25-e520-44a0-8948-5a48c706a332 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.077177] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 7350d352-9336-40b8-81a6-0a4795d9f8dd/7350d352-9336-40b8-81a6-0a4795d9f8dd.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.081445] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35b64dba-0e5d-4ccb-a073-5601c1c931fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.105712] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293931, 'name': ReconfigVM_Task, 'duration_secs': 0.487276} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.107456] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Reconfigured VM instance instance-0000001e to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 912.116022] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 912.116022] env[62814]: value = "task-4293935" [ 912.116022] env[62814]: _type = "Task" [ 912.116022] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.116022] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3128161-998b-4742-acc1-cbfd2a651144 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.141415] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for the task: (returnval){ [ 912.141415] env[62814]: value = "task-4293936" [ 912.141415] env[62814]: _type = "Task" [ 912.141415] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.151703] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293936, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.252024] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 912.252292] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 912.252476] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Deleting the datastore file [datastore2] cd89e4d5-c08e-42c9-aea8-84c20714e196 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 912.252873] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b285723f-d19e-444c-9ad2-ced298ff1316 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.260752] env[62814]: DEBUG oslo_vmware.api [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for the task: (returnval){ [ 912.260752] env[62814]: value = "task-4293937" [ 912.260752] env[62814]: _type = "Task" [ 912.260752] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.269749] env[62814]: DEBUG oslo_vmware.api [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293937, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.292650] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293923, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.479217] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 9677b03f-7138-47b7-b1e8-f3714d11e550] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 912.519939] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293933, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.526943] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293934, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.533277] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 912.533538] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 912.537507] env[62814]: INFO nova.compute.manager [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Attaching volume 252ca9ac-8b1f-4449-8f58-9fce53fe27ec to /dev/sdb [ 912.577016] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a92874c-793f-4c0f-aa9b-188205cad83e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.585372] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f26749-7614-44fe-99ed-dd55a9078283 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.600616] env[62814]: DEBUG nova.virt.block_device [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updating existing volume attachment record: 2fa25518-ef7b-42e5-b704-6a56608fdade {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 912.635896] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293935, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.645104] env[62814]: DEBUG nova.compute.manager [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 912.666035] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293936, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.699718] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 912.699718] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.699718] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 912.699916] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.700148] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 912.700300] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 912.700525] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 912.700726] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 912.700933] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 912.701146] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 912.701334] env[62814]: DEBUG nova.virt.hardware [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 912.702218] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a402aa-00ba-4367-b57f-ba2772b87634 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.715524] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af0bcd45-c47d-469a-81f2-6182ae17c3f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.772263] env[62814]: DEBUG oslo_vmware.api [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293937, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.790390] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293923, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.897828} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.793148] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/577d3b83-11e9-430b-a3da-3fc2f382fd79/577d3b83-11e9-430b-a3da-3fc2f382fd79.vmdk to [datastore2] af6ef867-dc9a-4db6-8582-a6198e5caa77/af6ef867-dc9a-4db6-8582-a6198e5caa77.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 912.794198] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c7ecce-fd1c-4069-8c07-d645aca413f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.826163] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] af6ef867-dc9a-4db6-8582-a6198e5caa77/af6ef867-dc9a-4db6-8582-a6198e5caa77.vmdk or device None with type streamOptimized {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 912.829326] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5eb14d6a-8576-4a8e-9bc2-7a20a68e2a43 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.851215] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 912.851215] env[62814]: value = "task-4293939" [ 912.851215] env[62814]: _type = "Task" [ 912.851215] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.863343] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293939, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.983532] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 1a16bd56-0992-4bec-bd3f-2836dfe68579] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 913.021342] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293933, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.026815] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293934, 'name': Rename_Task, 'duration_secs': 0.741125} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.027042] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 913.027292] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-70b9e810-e238-43a9-ae0a-f4667786f65d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.033257] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Waiting for the task: (returnval){ [ 913.033257] env[62814]: value = "task-4293941" [ 913.033257] env[62814]: _type = "Task" [ 913.033257] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.050672] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293941, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.142671] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293935, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.160644] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293936, 'name': ReconfigVM_Task, 'duration_secs': 0.931073} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.161145] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845587', 'volume_id': 'fa377baf-042f-48ba-8dc4-21baec93c237', 'name': 'volume-fa377baf-042f-48ba-8dc4-21baec93c237', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'bd0933ca-aab7-4dd4-a570-1a58a720f377', 'attached_at': '', 'detached_at': '', 'volume_id': 'fa377baf-042f-48ba-8dc4-21baec93c237', 'serial': 'fa377baf-042f-48ba-8dc4-21baec93c237'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 913.161418] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 913.164205] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d27a0a-2117-41fd-97ce-21b07d63bbd5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.172424] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 913.172739] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-807fdfd4-c1c7-4f52-9b75-68e20e84c7f7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.243256] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 913.243256] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 913.243869] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Deleting the datastore file [datastore2] bd0933ca-aab7-4dd4-a570-1a58a720f377 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 913.244360] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f979e79a-4fea-413b-8a73-33c89f9e487d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.254506] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for the task: (returnval){ [ 913.254506] env[62814]: value = "task-4293944" [ 913.254506] env[62814]: _type = "Task" [ 913.254506] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.260730] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f648087-607a-474e-9ce1-b7f49a59764b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.269342] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293944, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.281680] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc4a8fd-0594-4b37-894a-311078e7e6f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.285977] env[62814]: DEBUG oslo_vmware.api [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Task: {'id': task-4293937, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.548089} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.286341] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.286621] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 913.286889] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 913.287233] env[62814]: INFO nova.compute.manager [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Took 2.92 seconds to destroy the instance on the hypervisor. [ 913.287415] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 913.287989] env[62814]: DEBUG nova.compute.manager [-] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 913.288106] env[62814]: DEBUG nova.network.neutron [-] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 913.320073] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27527bf9-3f1d-4be5-9ac8-1bc6c5cc8b00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.328661] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ea0380-6ea0-4c91-ada0-ee7d6ede2401 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.343551] env[62814]: DEBUG nova.compute.provider_tree [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 913.361463] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293939, 'name': ReconfigVM_Task, 'duration_secs': 0.342163} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.361869] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Reconfigured VM instance instance-00000041 to attach disk [datastore2] af6ef867-dc9a-4db6-8582-a6198e5caa77/af6ef867-dc9a-4db6-8582-a6198e5caa77.vmdk or device None with type streamOptimized {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.362440] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-83c2c41f-f754-4b2b-84c1-1fe04d1cc77b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.368863] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 913.368863] env[62814]: value = "task-4293945" [ 913.368863] env[62814]: _type = "Task" [ 913.368863] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.378571] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293945, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.487966] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 56f9e03c-1157-442f-8add-156627a6ec1e] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 913.518520] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293933, 'name': ReconfigVM_Task, 'duration_secs': 1.386538} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.518849] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Reconfigured VM instance instance-00000040 to attach disk [datastore2] d7332882-19b3-4ab3-8ea6-51d33c584844/d7332882-19b3-4ab3-8ea6-51d33c584844.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.520456] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-905ea9aa-29e5-4a1b-bdab-59968dd8b23f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.528331] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 913.528331] env[62814]: value = "task-4293946" [ 913.528331] env[62814]: _type = "Task" [ 913.528331] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.539320] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293946, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.549638] env[62814]: DEBUG oslo_vmware.api [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Task: {'id': task-4293941, 'name': PowerOnVM_Task, 'duration_secs': 0.499553} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.550055] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 913.551029] env[62814]: DEBUG nova.compute.manager [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 913.552112] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb91a6d-1016-4e33-adc6-9a4defc5668a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.638121] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293935, 'name': ReconfigVM_Task, 'duration_secs': 1.243381} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.638541] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 7350d352-9336-40b8-81a6-0a4795d9f8dd/7350d352-9336-40b8-81a6-0a4795d9f8dd.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 913.639492] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ded52bdd-0746-4b63-8d29-f000d73f2745 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.650862] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 913.650862] env[62814]: value = "task-4293947" [ 913.650862] env[62814]: _type = "Task" [ 913.650862] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.656707] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293947, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.765686] env[62814]: DEBUG oslo_vmware.api [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Task: {'id': task-4293944, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.110717} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.766025] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 913.766222] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 913.766403] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 913.766579] env[62814]: INFO nova.compute.manager [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Took 2.88 seconds to destroy the instance on the hypervisor. [ 913.766820] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 913.767033] env[62814]: DEBUG nova.compute.manager [-] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 913.767144] env[62814]: DEBUG nova.network.neutron [-] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 913.805498] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 913.805757] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 913.879347] env[62814]: ERROR nova.scheduler.client.report [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [req-39d6e7b2-429d-4cbe-b6d0-952bf60168e9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-39d6e7b2-429d-4cbe-b6d0-952bf60168e9"}]} [ 913.890040] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293945, 'name': Rename_Task, 'duration_secs': 0.189108} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.890040] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 913.890040] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7253c0d-5e5b-4229-bfcf-11481062bce3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.894093] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 913.894093] env[62814]: value = "task-4293948" [ 913.894093] env[62814]: _type = "Task" [ 913.894093] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.902455] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293948, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.903915] env[62814]: DEBUG nova.scheduler.client.report [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 913.928765] env[62814]: DEBUG nova.scheduler.client.report [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 913.929095] env[62814]: DEBUG nova.compute.provider_tree [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 913.944810] env[62814]: DEBUG nova.scheduler.client.report [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 913.968341] env[62814]: DEBUG nova.scheduler.client.report [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 913.996206] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 961dedc6-5ddd-4620-bfa7-54cf0c520af3] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 914.037072] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293946, 'name': Rename_Task, 'duration_secs': 0.134681} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.039998] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.040451] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8063e96c-17d0-47ad-bdb6-f6cdb84986eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.047227] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 914.047227] env[62814]: value = "task-4293949" [ 914.047227] env[62814]: _type = "Task" [ 914.047227] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.056353] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293949, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.073273] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.159266] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293947, 'name': Rename_Task, 'duration_secs': 0.136039} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.159266] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.159266] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-147a0f21-67b0-4591-8228-72b2e5701d92 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.166839] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 914.166839] env[62814]: value = "task-4293950" [ 914.166839] env[62814]: _type = "Task" [ 914.166839] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.175093] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293950, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.249766] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "508321ab-ce10-4953-a9e3-193b9975bec7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.253018] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "508321ab-ce10-4953-a9e3-193b9975bec7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 914.257943] env[62814]: DEBUG nova.compute.manager [req-80d37682-8f25-45d1-b11b-fe06e32e8127 req-2dbee840-880d-4515-a331-4f9c1984f92e service nova] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Received event network-vif-plugged-5764c0a4-20c2-4baa-a42a-0968ba00981e {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 914.258176] env[62814]: DEBUG oslo_concurrency.lockutils [req-80d37682-8f25-45d1-b11b-fe06e32e8127 req-2dbee840-880d-4515-a331-4f9c1984f92e service nova] Acquiring lock "99864c68-5f11-4119-b8e3-3aa8719f267b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.258531] env[62814]: DEBUG oslo_concurrency.lockutils [req-80d37682-8f25-45d1-b11b-fe06e32e8127 req-2dbee840-880d-4515-a331-4f9c1984f92e service nova] Lock "99864c68-5f11-4119-b8e3-3aa8719f267b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 914.258613] env[62814]: DEBUG oslo_concurrency.lockutils [req-80d37682-8f25-45d1-b11b-fe06e32e8127 req-2dbee840-880d-4515-a331-4f9c1984f92e service nova] Lock "99864c68-5f11-4119-b8e3-3aa8719f267b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 914.258996] env[62814]: DEBUG nova.compute.manager [req-80d37682-8f25-45d1-b11b-fe06e32e8127 req-2dbee840-880d-4515-a331-4f9c1984f92e service nova] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] No waiting events found dispatching network-vif-plugged-5764c0a4-20c2-4baa-a42a-0968ba00981e {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 914.259893] env[62814]: WARNING nova.compute.manager [req-80d37682-8f25-45d1-b11b-fe06e32e8127 req-2dbee840-880d-4515-a331-4f9c1984f92e service nova] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Received unexpected event network-vif-plugged-5764c0a4-20c2-4baa-a42a-0968ba00981e for instance with vm_state building and task_state spawning. [ 914.307852] env[62814]: DEBUG nova.compute.manager [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 914.365427] env[62814]: DEBUG nova.network.neutron [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Successfully updated port: 5764c0a4-20c2-4baa-a42a-0968ba00981e {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 914.406855] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293948, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.503012] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 217e5812-95cc-4104-8d7b-82dfca2c0fcc] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 914.561414] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293949, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.585997] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6caa77d5-b457-433d-bbd4-be82bb9350ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.595677] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd42d9b-df4e-4d87-84d7-183cf2ea5067 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.634355] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bd7314-5457-4853-a579-f3025401d301 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.642981] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec9da3e-2cb0-4574-9093-6267e16a7b99 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.661397] env[62814]: DEBUG nova.compute.provider_tree [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 914.678898] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293950, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.738793] env[62814]: DEBUG nova.compute.manager [req-fdfa2f01-7d51-4efb-be7d-a1d9d050d68e req-2d248138-df33-4b80-9a86-1e929973d530 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Received event network-vif-deleted-e5baa36f-9802-41c1-add7-6020a8b74ce8 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 914.738991] env[62814]: INFO nova.compute.manager [req-fdfa2f01-7d51-4efb-be7d-a1d9d050d68e req-2d248138-df33-4b80-9a86-1e929973d530 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Neutron deleted interface e5baa36f-9802-41c1-add7-6020a8b74ce8; detaching it from the instance and deleting it from the info cache [ 914.739790] env[62814]: DEBUG nova.network.neutron [req-fdfa2f01-7d51-4efb-be7d-a1d9d050d68e req-2d248138-df33-4b80-9a86-1e929973d530 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.752576] env[62814]: DEBUG nova.compute.manager [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 914.835829] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.867129] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "refresh_cache-99864c68-5f11-4119-b8e3-3aa8719f267b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.867940] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired lock "refresh_cache-99864c68-5f11-4119-b8e3-3aa8719f267b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 914.867940] env[62814]: DEBUG nova.network.neutron [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.906396] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293948, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.991926] env[62814]: DEBUG nova.network.neutron [-] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.008788] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 369cd937-4c18-4068-ae59-70a1d585094b] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 915.029652] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "98884571-8101-4ae9-b2e3-aeed97e3618f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.030075] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "98884571-8101-4ae9-b2e3-aeed97e3618f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.030362] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "98884571-8101-4ae9-b2e3-aeed97e3618f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.030548] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "98884571-8101-4ae9-b2e3-aeed97e3618f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.030947] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "98884571-8101-4ae9-b2e3-aeed97e3618f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 915.034088] env[62814]: INFO nova.compute.manager [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Terminating instance [ 915.064356] env[62814]: DEBUG oslo_vmware.api [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293949, 'name': PowerOnVM_Task, 'duration_secs': 0.522345} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.064690] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.065201] env[62814]: DEBUG nova.compute.manager [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 915.066360] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f6e5cb4-d749-4a64-ae06-8495cfc68a20 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.180106] env[62814]: DEBUG oslo_vmware.api [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4293950, 'name': PowerOnVM_Task, 'duration_secs': 0.515515} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.180106] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.180106] env[62814]: INFO nova.compute.manager [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Took 11.94 seconds to spawn the instance on the hypervisor. [ 915.180106] env[62814]: DEBUG nova.compute.manager [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 915.181336] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca945384-a90e-430d-9613-d69dd3275594 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.198396] env[62814]: DEBUG nova.scheduler.client.report [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 95 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 915.198678] env[62814]: DEBUG nova.compute.provider_tree [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 95 to 96 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 915.198872] env[62814]: DEBUG nova.compute.provider_tree [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 915.219018] env[62814]: DEBUG nova.network.neutron [-] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.242315] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e78ed6b9-1bcc-4a21-a7c5-1d179ae734df {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.252062] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674d64a8-e8ca-4b75-9fb7-03303402822c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.293291] env[62814]: DEBUG nova.compute.manager [req-fdfa2f01-7d51-4efb-be7d-a1d9d050d68e req-2d248138-df33-4b80-9a86-1e929973d530 service nova] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Detach interface failed, port_id=e5baa36f-9802-41c1-add7-6020a8b74ce8, reason: Instance bd0933ca-aab7-4dd4-a570-1a58a720f377 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 915.294613] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.405913] env[62814]: DEBUG oslo_vmware.api [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293948, 'name': PowerOnVM_Task, 'duration_secs': 1.172362} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.406192] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.406617] env[62814]: INFO nova.compute.manager [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Took 16.95 seconds to spawn the instance on the hypervisor. [ 915.406812] env[62814]: DEBUG nova.compute.manager [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 915.409864] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74acfb08-66ff-4a9d-8228-aedcc1729806 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.479118] env[62814]: DEBUG nova.network.neutron [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 915.493754] env[62814]: INFO nova.compute.manager [-] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Took 2.21 seconds to deallocate network for instance. [ 915.513333] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: a589a3d8-20dc-4ff5-a192-c540e29f39d6] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 915.538315] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "refresh_cache-98884571-8101-4ae9-b2e3-aeed97e3618f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.538315] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquired lock "refresh_cache-98884571-8101-4ae9-b2e3-aeed97e3618f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 915.538813] env[62814]: DEBUG nova.network.neutron [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 915.584676] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 915.699933] env[62814]: INFO nova.compute.manager [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Took 45.36 seconds to build instance. [ 915.704504] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.082s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 915.705063] env[62814]: DEBUG nova.compute.manager [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 915.709504] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 29.900s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.718935] env[62814]: INFO nova.compute.manager [-] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Took 1.95 seconds to deallocate network for instance. [ 915.733716] env[62814]: DEBUG nova.network.neutron [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Updating instance_info_cache with network_info: [{"id": "5764c0a4-20c2-4baa-a42a-0968ba00981e", "address": "fa:16:3e:13:2d:ea", "network": {"id": "88bc6d73-6bc2-47d0-8e78-633cbe326db2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-765790142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558682c8ea0f4887874a1763f65cb9a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5764c0a4-20", "ovs_interfaceid": "5764c0a4-20c2-4baa-a42a-0968ba00981e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.933205] env[62814]: INFO nova.compute.manager [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Took 51.08 seconds to build instance. [ 916.001289] env[62814]: DEBUG oslo_concurrency.lockutils [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.021658] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: c7829a8b-a9f7-40b1-958c-732e2eaa2b3c] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 916.061933] env[62814]: DEBUG nova.network.neutron [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 916.135626] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "d7332882-19b3-4ab3-8ea6-51d33c584844" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.135626] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "d7332882-19b3-4ab3-8ea6-51d33c584844" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.135626] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "d7332882-19b3-4ab3-8ea6-51d33c584844-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 916.135626] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "d7332882-19b3-4ab3-8ea6-51d33c584844-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 916.135626] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "d7332882-19b3-4ab3-8ea6-51d33c584844-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.137541] env[62814]: INFO nova.compute.manager [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Terminating instance [ 916.158299] env[62814]: DEBUG nova.network.neutron [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.202132] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e77e6a64-8ecc-400b-946a-29979a661d35 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "7350d352-9336-40b8-81a6-0a4795d9f8dd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.060s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.212214] env[62814]: DEBUG nova.compute.utils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 916.213060] env[62814]: DEBUG nova.compute.manager [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Not allocating networking since 'none' was specified. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 916.216759] env[62814]: DEBUG nova.objects.instance [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lazy-loading 'migration_context' on Instance uuid 9cd89da4-d190-4bfc-81e1-da15b98d54ba {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.238669] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Releasing lock "refresh_cache-99864c68-5f11-4119-b8e3-3aa8719f267b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 916.242015] env[62814]: DEBUG nova.compute.manager [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Instance network_info: |[{"id": "5764c0a4-20c2-4baa-a42a-0968ba00981e", "address": "fa:16:3e:13:2d:ea", "network": {"id": "88bc6d73-6bc2-47d0-8e78-633cbe326db2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-765790142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558682c8ea0f4887874a1763f65cb9a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5764c0a4-20", "ovs_interfaceid": "5764c0a4-20c2-4baa-a42a-0968ba00981e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 916.242015] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:2d:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '209639b9-c313-4b35-86dc-dccd744d174a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5764c0a4-20c2-4baa-a42a-0968ba00981e', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 916.248048] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 916.249307] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 916.249706] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56fe513d-16f9-462a-aa56-27bbe18a9316 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.274533] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 916.274533] env[62814]: value = "task-4293952" [ 916.274533] env[62814]: _type = "Task" [ 916.274533] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.285636] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293952, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.290442] env[62814]: INFO nova.compute.manager [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Took 0.57 seconds to detach 1 volumes for instance. [ 916.290881] env[62814]: DEBUG nova.compute.manager [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Deleting volume: fa377baf-042f-48ba-8dc4-21baec93c237 {{(pid=62814) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3282}} [ 916.317690] env[62814]: DEBUG nova.compute.manager [req-0c8c62ab-29a3-44ec-acec-0a5e3512c3fb req-d4b59fcf-db46-426f-8a9e-951f967c4a03 service nova] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Received event network-changed-5764c0a4-20c2-4baa-a42a-0968ba00981e {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 916.317959] env[62814]: DEBUG nova.compute.manager [req-0c8c62ab-29a3-44ec-acec-0a5e3512c3fb req-d4b59fcf-db46-426f-8a9e-951f967c4a03 service nova] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Refreshing instance network info cache due to event network-changed-5764c0a4-20c2-4baa-a42a-0968ba00981e. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 916.318122] env[62814]: DEBUG oslo_concurrency.lockutils [req-0c8c62ab-29a3-44ec-acec-0a5e3512c3fb req-d4b59fcf-db46-426f-8a9e-951f967c4a03 service nova] Acquiring lock "refresh_cache-99864c68-5f11-4119-b8e3-3aa8719f267b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.318273] env[62814]: DEBUG oslo_concurrency.lockutils [req-0c8c62ab-29a3-44ec-acec-0a5e3512c3fb req-d4b59fcf-db46-426f-8a9e-951f967c4a03 service nova] Acquired lock "refresh_cache-99864c68-5f11-4119-b8e3-3aa8719f267b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 916.318433] env[62814]: DEBUG nova.network.neutron [req-0c8c62ab-29a3-44ec-acec-0a5e3512c3fb req-d4b59fcf-db46-426f-8a9e-951f967c4a03 service nova] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Refreshing network info cache for port 5764c0a4-20c2-4baa-a42a-0968ba00981e {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 916.436856] env[62814]: DEBUG oslo_concurrency.lockutils [None req-341e6cb3-7485-41ea-bf7e-5186dba0a4cd tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "af6ef867-dc9a-4db6-8582-a6198e5caa77" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.588s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 916.525593] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 004fd137-4902-4313-a6f7-6c83cd76743d] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 916.642172] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "refresh_cache-d7332882-19b3-4ab3-8ea6-51d33c584844" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.642632] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquired lock "refresh_cache-d7332882-19b3-4ab3-8ea6-51d33c584844" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 916.644541] env[62814]: DEBUG nova.network.neutron [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 916.661674] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Releasing lock "refresh_cache-98884571-8101-4ae9-b2e3-aeed97e3618f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 916.662440] env[62814]: DEBUG nova.compute.manager [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 916.662806] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 916.665469] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f0ec15-8018-4b15-ae78-a9c38dff3bb3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.677913] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 916.678546] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd1e3395-f489-4599-8550-09781f433123 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.688272] env[62814]: DEBUG oslo_vmware.api [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 916.688272] env[62814]: value = "task-4293954" [ 916.688272] env[62814]: _type = "Task" [ 916.688272] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.700816] env[62814]: DEBUG oslo_vmware.api [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293954, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.715033] env[62814]: DEBUG nova.compute.manager [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 916.788374] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293952, 'name': CreateVM_Task, 'duration_secs': 0.503261} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.791141] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 916.792225] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.792316] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 916.792615] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 916.792896] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-583aef1d-7e08-435f-8fc3-10be6540b4ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.798600] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 916.798600] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526b3eff-ffc3-592d-94e1-2e00e2180690" [ 916.798600] env[62814]: _type = "Task" [ 916.798600] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.812396] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526b3eff-ffc3-592d-94e1-2e00e2180690, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.855866] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.036274] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 961f313b-b43f-4531-8a4b-0a39421d6a34] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 917.036274] env[62814]: DEBUG oslo_concurrency.lockutils [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "af6ef867-dc9a-4db6-8582-a6198e5caa77" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.036274] env[62814]: DEBUG oslo_concurrency.lockutils [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "af6ef867-dc9a-4db6-8582-a6198e5caa77" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 917.036274] env[62814]: DEBUG oslo_concurrency.lockutils [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "af6ef867-dc9a-4db6-8582-a6198e5caa77-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.038165] env[62814]: DEBUG oslo_concurrency.lockutils [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "af6ef867-dc9a-4db6-8582-a6198e5caa77-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 917.038517] env[62814]: DEBUG oslo_concurrency.lockutils [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "af6ef867-dc9a-4db6-8582-a6198e5caa77-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.002s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 917.045242] env[62814]: INFO nova.compute.manager [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Terminating instance [ 917.171216] env[62814]: DEBUG nova.network.neutron [req-0c8c62ab-29a3-44ec-acec-0a5e3512c3fb req-d4b59fcf-db46-426f-8a9e-951f967c4a03 service nova] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Updated VIF entry in instance network info cache for port 5764c0a4-20c2-4baa-a42a-0968ba00981e. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 917.171573] env[62814]: DEBUG nova.network.neutron [req-0c8c62ab-29a3-44ec-acec-0a5e3512c3fb req-d4b59fcf-db46-426f-8a9e-951f967c4a03 service nova] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Updating instance_info_cache with network_info: [{"id": "5764c0a4-20c2-4baa-a42a-0968ba00981e", "address": "fa:16:3e:13:2d:ea", "network": {"id": "88bc6d73-6bc2-47d0-8e78-633cbe326db2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-765790142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "558682c8ea0f4887874a1763f65cb9a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "209639b9-c313-4b35-86dc-dccd744d174a", "external-id": "nsx-vlan-transportzone-868", "segmentation_id": 868, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5764c0a4-20", "ovs_interfaceid": "5764c0a4-20c2-4baa-a42a-0968ba00981e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.198423] env[62814]: DEBUG nova.network.neutron [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.208750] env[62814]: DEBUG oslo_vmware.api [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293954, 'name': PowerOffVM_Task, 'duration_secs': 0.160649} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.208750] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 917.208750] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 917.209036] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5a0759f2-c0f4-4902-97e5-ea2158a64279 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.242827] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 917.243096] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 917.243489] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Deleting the datastore file [datastore2] 98884571-8101-4ae9-b2e3-aeed97e3618f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 917.244402] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-47808f56-6339-495c-9259-2292aa386020 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.254281] env[62814]: DEBUG oslo_vmware.api [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for the task: (returnval){ [ 917.254281] env[62814]: value = "task-4293956" [ 917.254281] env[62814]: _type = "Task" [ 917.254281] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.269795] env[62814]: DEBUG oslo_vmware.api [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293956, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.295878] env[62814]: DEBUG nova.network.neutron [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.308665] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526b3eff-ffc3-592d-94e1-2e00e2180690, 'name': SearchDatastore_Task, 'duration_secs': 0.013889} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.309640] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 917.310452] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.310820] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.311011] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 917.311204] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.314450] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-838821e6-005d-4251-b5ae-7d89e48ead29 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.323486] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.324124] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 917.324490] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdca6d75-12e9-471f-9dc5-668f695e1ce7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.330333] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86fbd5f6-0a72-46c5-b70c-f9aaa061584d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.337404] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 917.337404] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e5bd89-19f3-1c1e-c20d-dd4d3cdd7218" [ 917.337404] env[62814]: _type = "Task" [ 917.337404] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.343391] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94eea9c9-d32a-4036-b338-f39754e57320 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.351789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "22429810-3ea7-4472-9a90-018d6efcba1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.351993] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "22429810-3ea7-4472-9a90-018d6efcba1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 917.389486] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e5bd89-19f3-1c1e-c20d-dd4d3cdd7218, 'name': SearchDatastore_Task, 'duration_secs': 0.011412} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.389486] env[62814]: DEBUG nova.compute.manager [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 917.389655] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322080c0-7694-4371-9d86-4b6aa6274708 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.393465] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15d22544-7885-46fa-a3a1-68e540eb23ee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.402536] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0853f9be-fdba-44b8-ae81-57d47019ebd7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.407385] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 917.407385] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527f3674-9594-c383-8ead-a403d65c34e8" [ 917.407385] env[62814]: _type = "Task" [ 917.407385] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.420026] env[62814]: DEBUG nova.compute.provider_tree [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 917.429021] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527f3674-9594-c383-8ead-a403d65c34e8, 'name': SearchDatastore_Task, 'duration_secs': 0.010378} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.429021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 917.429021] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 99864c68-5f11-4119-b8e3-3aa8719f267b/99864c68-5f11-4119-b8e3-3aa8719f267b.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 917.429021] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a4bc80bc-b061-4597-832b-ef11f5a54879 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.435721] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 917.435721] env[62814]: value = "task-4293957" [ 917.435721] env[62814]: _type = "Task" [ 917.435721] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.444695] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293957, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.545105] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 64a475e4-6713-408b-a63a-a43b5fed5ec8] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 917.552090] env[62814]: DEBUG nova.compute.manager [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 917.552090] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 917.554249] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8f9243-ac3f-4dff-92cd-2b2308de2d4e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.564734] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.565059] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5b53dac-ac42-48f3-9509-2ed21aa18639 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.571941] env[62814]: DEBUG oslo_vmware.api [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 917.571941] env[62814]: value = "task-4293958" [ 917.571941] env[62814]: _type = "Task" [ 917.571941] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.582267] env[62814]: DEBUG oslo_vmware.api [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293958, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.678130] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 917.678378] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845734', 'volume_id': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'name': 'volume-252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3e4479a0-8edd-4b37-8cc9-2c91275b88ee', 'attached_at': '', 'detached_at': '', 'volume_id': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'serial': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 917.678979] env[62814]: DEBUG oslo_concurrency.lockutils [req-0c8c62ab-29a3-44ec-acec-0a5e3512c3fb req-d4b59fcf-db46-426f-8a9e-951f967c4a03 service nova] Releasing lock "refresh_cache-99864c68-5f11-4119-b8e3-3aa8719f267b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 917.679226] env[62814]: DEBUG nova.compute.manager [req-0c8c62ab-29a3-44ec-acec-0a5e3512c3fb req-d4b59fcf-db46-426f-8a9e-951f967c4a03 service nova] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Received event network-vif-deleted-696f3530-9858-47e0-be99-e223f03b58fb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 917.680142] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1cf1e1-8ba7-4c80-ac92-615d420ca9ee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.698778] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0cac86-a302-4cd0-aced-02ad5934c54e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.725870] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Reconfiguring VM instance instance-0000002b to attach disk [datastore1] volume-252ca9ac-8b1f-4449-8f58-9fce53fe27ec/volume-252ca9ac-8b1f-4449-8f58-9fce53fe27ec.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.725870] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-527f815b-f368-455c-81c6-54269beb9ec8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.740847] env[62814]: DEBUG nova.compute.manager [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 917.750769] env[62814]: DEBUG oslo_vmware.api [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 917.750769] env[62814]: value = "task-4293959" [ 917.750769] env[62814]: _type = "Task" [ 917.750769] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.775542] env[62814]: DEBUG oslo_vmware.api [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293959, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.781974] env[62814]: DEBUG oslo_vmware.api [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Task: {'id': task-4293956, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142399} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.785767] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 917.786155] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.786444] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 917.786764] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.787056] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 917.787329] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 917.787717] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 917.788016] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 917.788387] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 917.788730] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 917.788956] env[62814]: DEBUG nova.virt.hardware [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 917.789410] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 917.789676] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 917.789959] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 917.790287] env[62814]: INFO nova.compute.manager [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 917.790717] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 917.792045] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8169d2ff-895e-4a9a-8ee0-2d4724948ded {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.796889] env[62814]: DEBUG nova.compute.manager [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 917.797106] env[62814]: DEBUG nova.network.neutron [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 917.800431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Releasing lock "refresh_cache-d7332882-19b3-4ab3-8ea6-51d33c584844" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 917.801121] env[62814]: DEBUG nova.compute.manager [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 917.801449] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 917.803527] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e6dd94-3109-4acd-9af3-faabd854bbc9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.816073] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-322fded1-5e79-4803-b138-8c1b5dc2d701 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.826049] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.827017] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-38b27556-debd-4315-bbce-7b18962b9247 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.844720] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 917.854981] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Creating folder: Project (433d7455290241f2bada1030217ce5ff). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.856364] env[62814]: DEBUG nova.network.neutron [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.859035] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7596baf-77ae-4e74-b462-1e01af6b31e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.869011] env[62814]: DEBUG oslo_vmware.api [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 917.869011] env[62814]: value = "task-4293960" [ 917.869011] env[62814]: _type = "Task" [ 917.869011] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.873376] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Created folder: Project (433d7455290241f2bada1030217ce5ff) in parent group-v845547. [ 917.873591] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Creating folder: Instances. Parent ref: group-v845736. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 917.874233] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83953db9-c398-4533-b180-439aa2144d43 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.879931] env[62814]: DEBUG oslo_vmware.api [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.888568] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Created folder: Instances in parent group-v845736. [ 917.889063] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 917.889281] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 917.889507] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-25321847-792b-4cec-a59f-187e453c55b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.915468] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 917.915468] env[62814]: value = "task-4293963" [ 917.915468] env[62814]: _type = "Task" [ 917.915468] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.927815] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293963, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.929468] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 917.952506] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293957, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.954446] env[62814]: ERROR nova.scheduler.client.report [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [req-5fd277d1-b8b8-4ceb-8788-e46b7860d64f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5fd277d1-b8b8-4ceb-8788-e46b7860d64f"}]} [ 917.977069] env[62814]: DEBUG nova.scheduler.client.report [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 917.993270] env[62814]: DEBUG nova.scheduler.client.report [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 917.993579] env[62814]: DEBUG nova.compute.provider_tree [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 918.014038] env[62814]: DEBUG nova.scheduler.client.report [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 918.046654] env[62814]: DEBUG nova.scheduler.client.report [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 918.053507] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: e6af4651-9f3a-4ce0-add8-06f1cfef255f] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 918.083900] env[62814]: DEBUG oslo_vmware.api [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293958, 'name': PowerOffVM_Task, 'duration_secs': 0.255733} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.084298] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.084487] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.085765] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc19866a-afb6-4854-aa2b-564174bdbdb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.110251] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 918.110602] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 918.147377] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 918.147934] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 918.147934] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleting the datastore file [datastore2] af6ef867-dc9a-4db6-8582-a6198e5caa77 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.148421] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2fb8ff35-49be-4124-b457-51f19d32e01f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.160799] env[62814]: DEBUG oslo_vmware.api [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 918.160799] env[62814]: value = "task-4293965" [ 918.160799] env[62814]: _type = "Task" [ 918.160799] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.178117] env[62814]: DEBUG oslo_vmware.api [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293965, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.262106] env[62814]: DEBUG oslo_vmware.api [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293959, 'name': ReconfigVM_Task, 'duration_secs': 0.492264} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.266074] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Reconfigured VM instance instance-0000002b to attach disk [datastore1] volume-252ca9ac-8b1f-4449-8f58-9fce53fe27ec/volume-252ca9ac-8b1f-4449-8f58-9fce53fe27ec.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.272034] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33e0e3ad-50a4-4721-aace-9d6b7cf24f38 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.289061] env[62814]: DEBUG oslo_vmware.api [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 918.289061] env[62814]: value = "task-4293966" [ 918.289061] env[62814]: _type = "Task" [ 918.289061] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.296435] env[62814]: DEBUG oslo_vmware.api [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293966, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.363085] env[62814]: DEBUG nova.network.neutron [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.381202] env[62814]: DEBUG oslo_vmware.api [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293960, 'name': PowerOffVM_Task, 'duration_secs': 0.367721} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.383998] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.383998] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.383998] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82319107-7a02-4c4b-8ef7-e4fd054e71a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.406656] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 918.406944] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 918.407172] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Deleting the datastore file [datastore2] d7332882-19b3-4ab3-8ea6-51d33c584844 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.410331] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-03c0a336-e721-4194-aafa-653d7045a339 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.423034] env[62814]: DEBUG oslo_vmware.api [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for the task: (returnval){ [ 918.423034] env[62814]: value = "task-4293968" [ 918.423034] env[62814]: _type = "Task" [ 918.423034] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.432632] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293963, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.437622] env[62814]: DEBUG oslo_vmware.api [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293968, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.449300] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293957, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.712118} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.449300] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 99864c68-5f11-4119-b8e3-3aa8719f267b/99864c68-5f11-4119-b8e3-3aa8719f267b.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.449300] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.449300] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4b785e3-445f-41a6-9d37-d68f34bd686b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.455459] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 918.455459] env[62814]: value = "task-4293969" [ 918.455459] env[62814]: _type = "Task" [ 918.455459] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.464373] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293969, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.557856] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 29e69c7c-08f7-4da4-9509-02a94bf971bc] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 918.613926] env[62814]: DEBUG nova.compute.manager [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 918.617937] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188a4b1e-d2c1-42d3-8832-62ab624b5dd3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.625236] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9cfa58f-f7b5-41e9-bf65-5d8255c48795 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.666268] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5131d2a-6c79-4c17-934d-8b49383d6146 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.674958] env[62814]: DEBUG oslo_vmware.api [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4293965, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.287132} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.677818] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.678111] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.678621] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.678808] env[62814]: INFO nova.compute.manager [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Took 1.13 seconds to destroy the instance on the hypervisor. [ 918.679126] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 918.679851] env[62814]: DEBUG nova.compute.manager [-] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 918.680019] env[62814]: DEBUG nova.network.neutron [-] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 918.682685] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-592c1339-8a60-457c-8e28-7e4c381502ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.698480] env[62814]: DEBUG nova.compute.provider_tree [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 918.797641] env[62814]: DEBUG oslo_vmware.api [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293966, 'name': ReconfigVM_Task, 'duration_secs': 0.147005} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.797972] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845734', 'volume_id': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'name': 'volume-252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3e4479a0-8edd-4b37-8cc9-2c91275b88ee', 'attached_at': '', 'detached_at': '', 'volume_id': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'serial': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 918.866627] env[62814]: INFO nova.compute.manager [-] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Took 1.07 seconds to deallocate network for instance. [ 918.937980] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293963, 'name': CreateVM_Task, 'duration_secs': 0.53598} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.939047] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 918.939047] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.939363] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 918.939944] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 918.943312] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9a14bad-b433-46fb-8304-d52d29657b64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.945342] env[62814]: DEBUG oslo_vmware.api [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Task: {'id': task-4293968, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153522} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.945726] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.945726] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.945897] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.946114] env[62814]: INFO nova.compute.manager [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Took 1.14 seconds to destroy the instance on the hypervisor. [ 918.948456] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 918.948456] env[62814]: DEBUG nova.compute.manager [-] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 918.948456] env[62814]: DEBUG nova.network.neutron [-] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 918.950565] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 918.950565] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef6843-2f46-63ff-cfdd-9c6abf9d9336" [ 918.950565] env[62814]: _type = "Task" [ 918.950565] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.962705] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef6843-2f46-63ff-cfdd-9c6abf9d9336, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.967600] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293969, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068268} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.968524] env[62814]: DEBUG nova.network.neutron [-] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 918.970216] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.974406] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33861f1-5dd4-4d91-98dd-b7f7cc8225a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.999763] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 99864c68-5f11-4119-b8e3-3aa8719f267b/99864c68-5f11-4119-b8e3-3aa8719f267b.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.999922] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0c5888b-45ab-448c-a7a1-d98045608c3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.020884] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 919.020884] env[62814]: value = "task-4293970" [ 919.020884] env[62814]: _type = "Task" [ 919.020884] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.029329] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293970, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.044229] env[62814]: DEBUG nova.compute.manager [req-d07c70e2-3d92-4e00-8771-e584f047fcde req-40567171-1112-4fb0-8a5c-98b688e0dd5e service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Received event network-vif-deleted-b8613625-0ec7-47e4-b3d3-8a9bef4be8c7 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 919.044430] env[62814]: INFO nova.compute.manager [req-d07c70e2-3d92-4e00-8771-e584f047fcde req-40567171-1112-4fb0-8a5c-98b688e0dd5e service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Neutron deleted interface b8613625-0ec7-47e4-b3d3-8a9bef4be8c7; detaching it from the instance and deleting it from the info cache [ 919.044598] env[62814]: DEBUG nova.network.neutron [req-d07c70e2-3d92-4e00-8771-e584f047fcde req-40567171-1112-4fb0-8a5c-98b688e0dd5e service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.062545] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 8213c50c-1d0c-4a4a-aae7-a5b1b0000fe4] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 919.138996] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.241398] env[62814]: DEBUG nova.scheduler.client.report [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 99 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 919.241645] env[62814]: DEBUG nova.compute.provider_tree [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 99 to 100 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 919.241813] env[62814]: DEBUG nova.compute.provider_tree [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 919.375948] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 919.462032] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef6843-2f46-63ff-cfdd-9c6abf9d9336, 'name': SearchDatastore_Task, 'duration_secs': 0.010052} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.462360] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 919.462591] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.462831] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.462979] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 919.463172] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.463421] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ebaa160-1afc-4629-bf6e-1ccc52e54322 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.471278] env[62814]: DEBUG nova.network.neutron [-] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.472868] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.473064] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 919.473749] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc8c8b4b-a578-4d6f-9219-bcef9d90fe16 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.478621] env[62814]: DEBUG nova.network.neutron [-] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.481795] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 919.481795] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ae178a-c475-897f-604a-8e5ae4f29f86" [ 919.481795] env[62814]: _type = "Task" [ 919.481795] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.489531] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ae178a-c475-897f-604a-8e5ae4f29f86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.532677] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293970, 'name': ReconfigVM_Task, 'duration_secs': 0.2852} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.533011] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 99864c68-5f11-4119-b8e3-3aa8719f267b/99864c68-5f11-4119-b8e3-3aa8719f267b.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 919.533643] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af504501-a4e5-4fce-8f80-bcf0950e0d44 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.539786] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 919.539786] env[62814]: value = "task-4293971" [ 919.539786] env[62814]: _type = "Task" [ 919.539786] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.547769] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293971, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.547989] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ce7bcfa-b990-4dc2-86be-c9b4e63b9aa8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.556688] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302db8a8-8f2e-40f8-8451-19d8ac4d9232 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.567295] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 8e04a5b7-7cbf-46ac-b5a0-bc21b5e4209e] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 919.596101] env[62814]: DEBUG nova.compute.manager [req-d07c70e2-3d92-4e00-8771-e584f047fcde req-40567171-1112-4fb0-8a5c-98b688e0dd5e service nova] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Detach interface failed, port_id=b8613625-0ec7-47e4-b3d3-8a9bef4be8c7, reason: Instance af6ef867-dc9a-4db6-8582-a6198e5caa77 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 919.841023] env[62814]: DEBUG nova.objects.instance [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lazy-loading 'flavor' on Instance uuid 3e4479a0-8edd-4b37-8cc9-2c91275b88ee {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.974200] env[62814]: INFO nova.compute.manager [-] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Took 1.03 seconds to deallocate network for instance. [ 919.984484] env[62814]: INFO nova.compute.manager [-] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Took 1.30 seconds to deallocate network for instance. [ 919.998506] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ae178a-c475-897f-604a-8e5ae4f29f86, 'name': SearchDatastore_Task, 'duration_secs': 0.010004} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.999359] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-766e0b35-7e40-45a2-a18a-14a483179863 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.004833] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 920.004833] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524330f7-39ec-da90-4c43-4571856d2d99" [ 920.004833] env[62814]: _type = "Task" [ 920.004833] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.013463] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524330f7-39ec-da90-4c43-4571856d2d99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.049582] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293971, 'name': Rename_Task, 'duration_secs': 0.144608} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.049894] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.050292] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-35a3eee5-c7b5-4a0f-8b42-460c95339dd9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.056083] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 920.056083] env[62814]: value = "task-4293972" [ 920.056083] env[62814]: _type = "Task" [ 920.056083] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.064032] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293972, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.071576] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: a36ad785-2f33-4dbc-bc82-ab4a35020b0f] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 920.101600] env[62814]: INFO nova.compute.manager [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Rescuing [ 920.102171] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.102729] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 920.102729] env[62814]: DEBUG nova.network.neutron [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.254609] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 4.545s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.260429] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 33.919s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 920.261098] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.262715] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.041s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 920.262932] env[62814]: DEBUG nova.objects.instance [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 920.289957] env[62814]: INFO nova.scheduler.client.report [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleted allocations for instance ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01 [ 920.349327] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ef9b8eb-c63a-477e-aa3f-a3a648f65361 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.816s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.481843] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.495033] env[62814]: DEBUG oslo_concurrency.lockutils [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 920.515598] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524330f7-39ec-da90-4c43-4571856d2d99, 'name': SearchDatastore_Task, 'duration_secs': 0.010297} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.515914] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 920.516199] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a/0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.516515] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e3e55f54-7d2b-40d7-bf46-2ccfe60657c7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.523199] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 920.523199] env[62814]: value = "task-4293973" [ 920.523199] env[62814]: _type = "Task" [ 920.523199] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.530935] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293973, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.565483] env[62814]: DEBUG oslo_vmware.api [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293972, 'name': PowerOnVM_Task, 'duration_secs': 0.433437} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.565809] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 920.566048] env[62814]: INFO nova.compute.manager [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Took 7.92 seconds to spawn the instance on the hypervisor. [ 920.566233] env[62814]: DEBUG nova.compute.manager [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 920.566994] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb3bceb-edea-4ca1-97e0-93e8703bf8dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.576321] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 98d134b4-b4ca-4247-a638-ad5c24a694e5] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 920.805183] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7946b498-f8cd-42de-b60c-39a626654b43 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.023s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 920.931013] env[62814]: DEBUG nova.network.neutron [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updating instance_info_cache with network_info: [{"id": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "address": "fa:16:3e:aa:55:02", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b486fb-bb", "ovs_interfaceid": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.034222] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293973, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505266} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.034494] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a/0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 921.034714] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 921.034974] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a9f9a55b-715c-4b8b-a13f-e6048f06502f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.041383] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 921.041383] env[62814]: value = "task-4293974" [ 921.041383] env[62814]: _type = "Task" [ 921.041383] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.049741] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293974, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.081053] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 296f6c11-7108-42e6-8ada-5d8c08b00da6] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 921.086148] env[62814]: INFO nova.compute.manager [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Took 41.04 seconds to build instance. [ 921.120738] env[62814]: DEBUG oslo_concurrency.lockutils [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.120738] env[62814]: DEBUG oslo_concurrency.lockutils [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.121187] env[62814]: DEBUG oslo_concurrency.lockutils [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 921.121409] env[62814]: DEBUG oslo_concurrency.lockutils [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.121585] env[62814]: DEBUG oslo_concurrency.lockutils [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.123502] env[62814]: INFO nova.compute.manager [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Terminating instance [ 921.281582] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4ec1e754-8bf4-4378-88ea-c9cee3fcf6a0 tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 921.282693] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.915s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 921.284951] env[62814]: INFO nova.compute.claims [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.434476] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 922.213735] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 6976b964-a8d3-4886-8aac-8d513e721018] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 922.215554] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cd44aba7-c14c-49e0-9099-ffea6156cc46 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "99864c68-5f11-4119-b8e3-3aa8719f267b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.185s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 922.216075] env[62814]: DEBUG nova.compute.manager [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 922.216266] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 922.229464] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a086234d-d3f4-406f-a6b5-13d8781a7e0e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.242028] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293974, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.190966} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.243850] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 922.244404] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 922.246033] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0449c98a-bd12-4774-8647-c820a9a42ec8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.249318] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e034e64-b47a-43ed-8df8-413957480c83 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.252455] env[62814]: INFO nova.compute.manager [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Swapping old allocation on dict_keys(['7136a6f7-3927-4aa1-a4e9-7fcbd9976745']) held by migration 8d6e5dee-326f-4a97-a651-a24c301bccc8 for instance [ 922.273357] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a/0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 922.278468] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22a57451-1626-4b15-b4e4-2c48369fcdec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.292979] env[62814]: DEBUG oslo_vmware.api [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 922.292979] env[62814]: value = "task-4293975" [ 922.292979] env[62814]: _type = "Task" [ 922.292979] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.298872] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 922.298872] env[62814]: value = "task-4293976" [ 922.298872] env[62814]: _type = "Task" [ 922.298872] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.304924] env[62814]: DEBUG oslo_vmware.api [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293975, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.305828] env[62814]: DEBUG nova.scheduler.client.report [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Overwriting current allocation {'allocations': {'7136a6f7-3927-4aa1-a4e9-7fcbd9976745': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 100}}, 'project_id': '8003fcc64a3147a296244034cfd77c9f', 'user_id': 'd147453e63fd4439829b3502a3d6db97', 'consumer_generation': 1} on consumer 9cd89da4-d190-4bfc-81e1-da15b98d54ba {{(pid=62814) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 922.313557] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293976, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.411143] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.411334] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquired lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 922.411517] env[62814]: DEBUG nova.network.neutron [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.729686] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 3d3f32fc-276a-49be-b471-01a5d6fc5069] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 922.755895] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "99864c68-5f11-4119-b8e3-3aa8719f267b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.756177] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "99864c68-5f11-4119-b8e3-3aa8719f267b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.756385] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "99864c68-5f11-4119-b8e3-3aa8719f267b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 922.756563] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "99864c68-5f11-4119-b8e3-3aa8719f267b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 922.756731] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "99864c68-5f11-4119-b8e3-3aa8719f267b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 922.761271] env[62814]: INFO nova.compute.manager [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Terminating instance [ 922.809680] env[62814]: DEBUG oslo_vmware.api [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293975, 'name': PowerOffVM_Task, 'duration_secs': 0.178994} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.814838] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 922.815052] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 922.815292] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293976, 'name': ReconfigVM_Task, 'duration_secs': 0.324932} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.815808] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa0641ad-4359-42cd-946a-4ec6ace0c4d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.817107] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a/0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.817662] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7a8aad4-e795-42e1-b7bf-d5b19e57fd66 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.824360] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 922.824360] env[62814]: value = "task-4293978" [ 922.824360] env[62814]: _type = "Task" [ 922.824360] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.836195] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293978, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.889541] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 922.889779] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 922.889927] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleting the datastore file [datastore2] cbc5985a-38e7-4e52-9fb0-264b5cec013a {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 922.890198] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c70db37c-add2-48f1-8483-322ff482b005 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.896844] env[62814]: DEBUG oslo_vmware.api [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 922.896844] env[62814]: value = "task-4293979" [ 922.896844] env[62814]: _type = "Task" [ 922.896844] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.908524] env[62814]: DEBUG oslo_vmware.api [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293979, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.212284] env[62814]: DEBUG nova.network.neutron [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance_info_cache with network_info: [{"id": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "address": "fa:16:3e:fe:5b:6c", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.210", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc38976c-88", "ovs_interfaceid": "bc38976c-8838-47e6-b4f8-cd013e62b15b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.233481] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 5087b202-9bba-4489-823b-5d93cbf116e2] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 923.235878] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.235878] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-266cbb53-08bd-4894-af02-b12086e95b77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.240038] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1846b47-b700-4ea0-8a10-11664fde076a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.243452] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 923.243452] env[62814]: value = "task-4293980" [ 923.243452] env[62814]: _type = "Task" [ 923.243452] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.250483] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25a0b06-d4a1-452e-a9b4-f65d115fd6ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.257068] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293980, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.285405] env[62814]: DEBUG nova.compute.manager [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 923.285623] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 923.287609] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c7f786-47c2-4e9a-8f77-08effb134507 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.290806] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3be1724-ec0a-4166-800c-6c99e3bdd716 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.300563] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3801e3b0-4f35-4883-b751-d627bf3d03d1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.305038] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.305299] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-98cd744f-ebdf-463d-a42f-7046c463417e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.319369] env[62814]: DEBUG nova.compute.provider_tree [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.323103] env[62814]: DEBUG oslo_vmware.api [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 923.323103] env[62814]: value = "task-4293981" [ 923.323103] env[62814]: _type = "Task" [ 923.323103] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.337037] env[62814]: DEBUG oslo_vmware.api [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293981, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.339063] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293978, 'name': Rename_Task, 'duration_secs': 0.153034} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.339363] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 923.339612] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-45a9e53c-b343-4570-808d-06c632111993 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.346531] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 923.346531] env[62814]: value = "task-4293982" [ 923.346531] env[62814]: _type = "Task" [ 923.346531] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.355122] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293982, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.410632] env[62814]: DEBUG oslo_vmware.api [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4293979, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145464} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.410861] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.411079] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 923.411259] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 923.411449] env[62814]: INFO nova.compute.manager [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Took 1.20 seconds to destroy the instance on the hypervisor. [ 923.411756] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 923.411982] env[62814]: DEBUG nova.compute.manager [-] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 923.412090] env[62814]: DEBUG nova.network.neutron [-] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 923.715340] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Releasing lock "refresh_cache-9cd89da4-d190-4bfc-81e1-da15b98d54ba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 923.717679] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.720499] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d789e97-b5d0-4481-b614-3fe422e69970 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.728019] env[62814]: DEBUG nova.compute.manager [req-7f1894da-4d27-47f5-bc09-d1f3b6576037 req-00a1ebba-0c0a-4730-b603-c3e2148eba92 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Received event network-vif-deleted-1b565819-b3b3-4e33-b334-07d6fd8b9d62 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 923.728019] env[62814]: INFO nova.compute.manager [req-7f1894da-4d27-47f5-bc09-d1f3b6576037 req-00a1ebba-0c0a-4730-b603-c3e2148eba92 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Neutron deleted interface 1b565819-b3b3-4e33-b334-07d6fd8b9d62; detaching it from the instance and deleting it from the info cache [ 923.728019] env[62814]: DEBUG nova.network.neutron [req-7f1894da-4d27-47f5-bc09-d1f3b6576037 req-00a1ebba-0c0a-4730-b603-c3e2148eba92 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.732028] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 923.732028] env[62814]: value = "task-4293983" [ 923.732028] env[62814]: _type = "Task" [ 923.732028] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.738683] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 62bc755d-4f96-4486-884b-0d0c337267aa] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 923.744477] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293983, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.753798] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293980, 'name': PowerOffVM_Task, 'duration_secs': 0.257084} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.753883] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 923.754809] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71720b0-12be-4dd2-9e62-8067fe80151f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.777418] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd1ef21-be33-4183-a768-16741f19b7da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.825113] env[62814]: DEBUG nova.scheduler.client.report [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 923.830760] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.834038] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cbef8060-f4c8-470e-8369-29c5c57a6b10 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.841262] env[62814]: DEBUG oslo_vmware.api [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293981, 'name': PowerOffVM_Task, 'duration_secs': 0.187856} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.842550] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 923.842727] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 923.843052] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 923.843052] env[62814]: value = "task-4293984" [ 923.843052] env[62814]: _type = "Task" [ 923.843052] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.843829] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0c21f2a-647d-4cda-a008-4f925cf2b2b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.858694] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293982, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.862296] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 923.862480] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.862737] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.862885] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 923.863086] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.863643] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad703c2f-0c2a-412c-9c08-f0942e42b0f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.873313] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.874390] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.874614] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6bb7fcc-6a47-4493-8895-06d55191d5db {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.880928] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 923.880928] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aacedd-ee82-e60f-3dcc-4838706be6d6" [ 923.880928] env[62814]: _type = "Task" [ 923.880928] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.889224] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aacedd-ee82-e60f-3dcc-4838706be6d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.912068] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 923.912068] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 923.912068] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Deleting the datastore file [datastore2] 99864c68-5f11-4119-b8e3-3aa8719f267b {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 923.912068] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9358b754-8c2b-42ae-b25f-c01bb2fea94c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.916958] env[62814]: DEBUG oslo_vmware.api [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for the task: (returnval){ [ 923.916958] env[62814]: value = "task-4293986" [ 923.916958] env[62814]: _type = "Task" [ 923.916958] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.925334] env[62814]: DEBUG oslo_vmware.api [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293986, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.174086] env[62814]: DEBUG nova.network.neutron [-] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.228422] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b51c0de-738c-4cfc-8654-688a9d6dc032 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.240388] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbbcc4c6-2d95-41b3-8eeb-62ad66cb82f2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.254160] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 17bfe703-ff96-4cb7-8535-eed57e10d673] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 924.256289] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293983, 'name': PowerOffVM_Task, 'duration_secs': 0.18971} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.257108] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.257735] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:36:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='82139a72-836d-436a-81f9-b33cf43b83c7',id=40,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1395970101',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 924.257944] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.258108] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 924.258290] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.258439] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 924.258628] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 924.258839] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 924.259009] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 924.259175] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 924.259333] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 924.259499] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 924.265153] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51deb512-f538-4561-9485-9940db8fee86 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.288423] env[62814]: DEBUG nova.compute.manager [req-7f1894da-4d27-47f5-bc09-d1f3b6576037 req-00a1ebba-0c0a-4730-b603-c3e2148eba92 service nova] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Detach interface failed, port_id=1b565819-b3b3-4e33-b334-07d6fd8b9d62, reason: Instance cbc5985a-38e7-4e52-9fb0-264b5cec013a could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 924.295288] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 924.295288] env[62814]: value = "task-4293987" [ 924.295288] env[62814]: _type = "Task" [ 924.295288] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.303742] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.336692] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.054s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 924.337324] env[62814]: DEBUG nova.compute.manager [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 924.339941] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.789s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 924.340633] env[62814]: DEBUG nova.objects.instance [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lazy-loading 'resources' on Instance uuid 310e22c3-ff0a-4e6b-aa46-8812fa014dfa {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.359373] env[62814]: DEBUG oslo_vmware.api [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4293982, 'name': PowerOnVM_Task, 'duration_secs': 0.770819} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.359638] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 924.359834] env[62814]: INFO nova.compute.manager [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Took 6.62 seconds to spawn the instance on the hypervisor. [ 924.360014] env[62814]: DEBUG nova.compute.manager [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 924.360801] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f03eef-4dcd-47d1-b59a-a484fe361209 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.391714] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aacedd-ee82-e60f-3dcc-4838706be6d6, 'name': SearchDatastore_Task, 'duration_secs': 0.01203} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.392563] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fcb047e-21f9-488d-a87d-60e0054ce36e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.397711] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 924.397711] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7cd60-541d-cf30-09f7-cf4b81d3cccc" [ 924.397711] env[62814]: _type = "Task" [ 924.397711] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.405669] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7cd60-541d-cf30-09f7-cf4b81d3cccc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.426333] env[62814]: DEBUG oslo_vmware.api [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Task: {'id': task-4293986, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.331652} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.426624] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.426825] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 924.427042] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 924.427220] env[62814]: INFO nova.compute.manager [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 924.427450] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 924.427634] env[62814]: DEBUG nova.compute.manager [-] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 924.427727] env[62814]: DEBUG nova.network.neutron [-] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 924.676484] env[62814]: INFO nova.compute.manager [-] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Took 1.26 seconds to deallocate network for instance. [ 924.759644] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 0ebb226a-40a3-40cd-8492-ccce87baf0e6] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 924.804536] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293987, 'name': ReconfigVM_Task, 'duration_secs': 0.342725} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.805322] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa4eb30-76b6-4a37-ab36-cb5407d1d8c7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.825213] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:36:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='82139a72-836d-436a-81f9-b33cf43b83c7',id=40,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1395970101',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 924.825560] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.825730] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 924.825950] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.826143] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 924.826323] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 924.826565] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 924.826764] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 924.827046] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 924.827153] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 924.827328] env[62814]: DEBUG nova.virt.hardware [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 924.828488] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f267c08-c111-4c87-9aab-f11c7fc35414 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.834039] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 924.834039] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dff9c2-457f-8ae8-b3f6-03eda5dddafe" [ 924.834039] env[62814]: _type = "Task" [ 924.834039] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.843547] env[62814]: DEBUG nova.compute.utils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 924.847681] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dff9c2-457f-8ae8-b3f6-03eda5dddafe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.848154] env[62814]: DEBUG nova.compute.manager [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 924.848323] env[62814]: DEBUG nova.network.neutron [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 924.876131] env[62814]: INFO nova.compute.manager [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Took 39.63 seconds to build instance. [ 924.878148] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "1e259ec6-d31a-453b-87e0-baa446665d56" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 924.878370] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 924.896928] env[62814]: DEBUG nova.policy [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b96cac63d9f47e8942a0e9e3e9c96bd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd5f0190bd5f24a909c9839b7fd1e870b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 924.910431] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7cd60-541d-cf30-09f7-cf4b81d3cccc, 'name': SearchDatastore_Task, 'duration_secs': 0.013994} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.910788] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 924.911083] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 3e4479a0-8edd-4b37-8cc9-2c91275b88ee/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. {{(pid=62814) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 924.911347] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8ae7b12b-4040-4759-813d-d61979498c3f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.920049] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 924.920049] env[62814]: value = "task-4293988" [ 924.920049] env[62814]: _type = "Task" [ 924.920049] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.933047] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293988, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.171440] env[62814]: DEBUG nova.network.neutron [-] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.183230] env[62814]: DEBUG oslo_concurrency.lockutils [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 925.255179] env[62814]: DEBUG nova.network.neutron [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Successfully created port: f3e3a6f1-ae61-48eb-a135-02afe3f22971 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.262828] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: f5504590-823d-4c37-8226-26684f7957a6] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 925.344103] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dff9c2-457f-8ae8-b3f6-03eda5dddafe, 'name': SearchDatastore_Task, 'duration_secs': 0.013904} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.349627] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Reconfiguring VM instance instance-00000030 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 925.351264] env[62814]: DEBUG nova.compute.manager [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 925.353978] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5186bbcd-65c2-44a9-934b-a79e2bb3b959 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.369313] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80bf5739-e697-40b7-bea5-759c6ddc0d7d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.378049] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d63234ae-f7a0-440f-a7c2-22eccbf45749 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.156s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.383323] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b14eb0-a411-47e5-a359-20f9c6094b97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.385289] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 925.385289] env[62814]: value = "task-4293989" [ 925.385289] env[62814]: _type = "Task" [ 925.385289] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.386285] env[62814]: DEBUG nova.compute.utils [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 925.422363] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5163de-d688-47e4-8f23-9490ad14c7bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.432234] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.441425] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f8dab1-7faa-4980-9110-79472c3e6cfe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.448315] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293988, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.458882] env[62814]: DEBUG nova.compute.provider_tree [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.674612] env[62814]: INFO nova.compute.manager [-] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Took 1.25 seconds to deallocate network for instance. [ 925.768128] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: be02bc93-1278-4b3d-afa3-270c84585d1d] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 925.898284] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293989, 'name': ReconfigVM_Task, 'duration_secs': 0.313918} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.898284] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Reconfigured VM instance instance-00000030 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 925.898284] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2e21fb-b911-48c9-9950-bbf9aa324539 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.922168] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 9cd89da4-d190-4bfc-81e1-da15b98d54ba/9cd89da4-d190-4bfc-81e1-da15b98d54ba.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.923741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.045s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 925.924251] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a538c8a-bab3-45c7-9c72-8daf41ef8b2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.949177] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293988, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.71857} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.950533] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 3e4479a0-8edd-4b37-8cc9-2c91275b88ee/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. [ 925.950971] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 925.950971] env[62814]: value = "task-4293990" [ 925.950971] env[62814]: _type = "Task" [ 925.950971] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.952025] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9fd34e-5873-40c3-b5f1-680fed599da7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.978876] env[62814]: DEBUG nova.scheduler.client.report [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 925.984292] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293990, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.992113] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 3e4479a0-8edd-4b37-8cc9-2c91275b88ee/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.992623] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-128fc62b-1d24-459f-a9de-af42516230bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.011773] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 926.011773] env[62814]: value = "task-4293991" [ 926.011773] env[62814]: _type = "Task" [ 926.011773] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.020095] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293991, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.181348] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 926.271675] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 66db614c-d101-409d-a5c3-9de38e5a6eff] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 926.384405] env[62814]: DEBUG nova.compute.manager [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 926.408966] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 926.409247] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.410158] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 926.410610] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.410848] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 926.411055] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 926.411279] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 926.411458] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 926.411632] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 926.411810] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 926.412042] env[62814]: DEBUG nova.virt.hardware [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 926.412919] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08ccba4-753a-401a-91ae-08ca293396e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.421115] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d006a6ed-530d-4f94-9664-7b21435e7ab9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.463399] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293990, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.486040] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.146s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 926.488837] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.949s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.488837] env[62814]: DEBUG nova.objects.instance [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lazy-loading 'resources' on Instance uuid 1f40dc62-1a58-4cfb-8785-c37b68747f37 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.513522] env[62814]: INFO nova.scheduler.client.report [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Deleted allocations for instance 310e22c3-ff0a-4e6b-aa46-8812fa014dfa [ 926.521270] env[62814]: DEBUG nova.compute.manager [req-6d8092e3-a8f9-4557-bdce-b8999291fef0 req-a25cf7c7-cca6-43d8-9592-fb063e9538d2 service nova] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Received event network-vif-deleted-5764c0a4-20c2-4baa-a42a-0968ba00981e {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 926.530717] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293991, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.776030] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 591e4a92-7c03-4d70-885f-49a66032c97e] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 926.815078] env[62814]: DEBUG nova.compute.manager [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 926.816063] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de60fc0-c0c0-40d1-95b5-12dba7e84d83 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.952039] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "1e259ec6-d31a-453b-87e0-baa446665d56" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 926.952425] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 926.952678] env[62814]: INFO nova.compute.manager [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Attaching volume 4bf78821-2028-4edc-a038-062cb9d09c71 to /dev/sdb [ 926.964566] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293990, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.015900] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6761181-f9b6-4559-a687-ae96d7202cfd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.033581] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293991, 'name': ReconfigVM_Task, 'duration_secs': 1.010788} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.034655] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9242f4-f85c-449a-b210-6f925b804fe9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.039066] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 3e4479a0-8edd-4b37-8cc9-2c91275b88ee/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.040110] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90006d9b-f963-4d0e-908c-149a3a5eaac2 tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "310e22c3-ff0a-4e6b-aa46-8812fa014dfa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.090s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 927.041153] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc68f2b-1eba-413c-b585-d7270e35302f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.075751] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e65eed4-054c-4c2e-b61a-e2a5558e8882 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.086436] env[62814]: DEBUG nova.virt.block_device [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Updating existing volume attachment record: 1001109b-8cc1-48da-b3e3-dcd4ba8eb3d0 {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 927.092299] env[62814]: DEBUG nova.network.neutron [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Successfully updated port: f3e3a6f1-ae61-48eb-a135-02afe3f22971 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 927.096369] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 927.096369] env[62814]: value = "task-4293992" [ 927.096369] env[62814]: _type = "Task" [ 927.096369] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.105255] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293992, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.283943] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 37649169-66e6-4ef1-a4cd-c65b9831004a] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 927.327976] env[62814]: INFO nova.compute.manager [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] instance snapshotting [ 927.328666] env[62814]: DEBUG nova.objects.instance [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'flavor' on Instance uuid 9343301c-dfe9-41b0-b4a0-067af544d297 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.464926] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293990, 'name': ReconfigVM_Task, 'duration_secs': 1.036338} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.467516] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 9cd89da4-d190-4bfc-81e1-da15b98d54ba/9cd89da4-d190-4bfc-81e1-da15b98d54ba.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.469733] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a06149a-b2f5-4768-b71f-6eca722b5842 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.497654] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c9244be-34cc-4588-869a-9c8c7a850510 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.523392] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aa023a9-b418-4c6c-b186-1d1c47b87203 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.548267] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c5d98d-28e6-47c0-b419-3448c86c87e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.555970] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.556264] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49a27b07-b0eb-4e47-b2e1-81562d440373 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.562639] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 927.562639] env[62814]: value = "task-4293996" [ 927.562639] env[62814]: _type = "Task" [ 927.562639] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.576236] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293996, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.579996] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d646f8f3-94a3-4eff-9e5d-1275851a0cf5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.586777] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2464d7a4-0a0b-45ea-82b6-959b4ed68841 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.618435] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "refresh_cache-5c574786-64a4-4f07-a267-101ecaaa6938" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.618604] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquired lock "refresh_cache-5c574786-64a4-4f07-a267-101ecaaa6938" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 927.618780] env[62814]: DEBUG nova.network.neutron [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 927.624335] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77f7400-afbb-4649-a585-1778c8c5edce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.632412] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293992, 'name': ReconfigVM_Task, 'duration_secs': 0.162571} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.634626] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.635677] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4f51703-1c32-4cdd-b87b-bc01b83cbfee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.637698] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d3d2913-b16b-4050-9634-747cd19dffef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.654751] env[62814]: DEBUG nova.compute.provider_tree [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 927.657170] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 927.657170] env[62814]: value = "task-4293997" [ 927.657170] env[62814]: _type = "Task" [ 927.657170] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.665437] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293997, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.787862] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 32f1bd48-3a2e-42fe-8033-85d3ad55affe] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 927.834981] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e7b514-9264-433a-a864-2db12a828806 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.854697] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6b7f90-e39d-4e52-b8fc-9df42d110682 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.886298] env[62814]: INFO nova.compute.manager [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Rebuilding instance [ 927.934992] env[62814]: DEBUG nova.compute.manager [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 927.936117] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1066f339-fc27-47da-a3cc-c2ea7d05eb44 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.077505] env[62814]: DEBUG oslo_vmware.api [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4293996, 'name': PowerOnVM_Task, 'duration_secs': 0.450603} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.077898] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.174689] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293997, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.180281] env[62814]: DEBUG nova.network.neutron [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 928.184765] env[62814]: ERROR nova.scheduler.client.report [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] [req-468996cf-9339-456f-9c86-1c9b68551194] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-468996cf-9339-456f-9c86-1c9b68551194"}]} [ 928.201078] env[62814]: DEBUG nova.scheduler.client.report [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 928.237017] env[62814]: DEBUG nova.scheduler.client.report [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 928.237392] env[62814]: DEBUG nova.compute.provider_tree [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 928.255382] env[62814]: DEBUG nova.scheduler.client.report [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 928.277253] env[62814]: DEBUG nova.scheduler.client.report [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 928.294070] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 928.294369] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Cleaning up deleted instances with incomplete migration {{(pid=62814) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11903}} [ 928.366185] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 928.366651] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-fc27a3d1-3172-41e5-b22f-c4bb55b8a6c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.376435] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 928.376435] env[62814]: value = "task-4293998" [ 928.376435] env[62814]: _type = "Task" [ 928.376435] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.389278] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293998, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.409135] env[62814]: DEBUG nova.network.neutron [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Updating instance_info_cache with network_info: [{"id": "f3e3a6f1-ae61-48eb-a135-02afe3f22971", "address": "fa:16:3e:f9:d2:13", "network": {"id": "b16d5204-f140-44fb-b4c4-621c3f5a3423", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-309109771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5f0190bd5f24a909c9839b7fd1e870b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec46b14d-3310-4f2b-96c1-f53ee47d3759", "external-id": "nsx-vlan-transportzone-323", "segmentation_id": 323, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3e3a6f1-ae", "ovs_interfaceid": "f3e3a6f1-ae61-48eb-a135-02afe3f22971", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.535955] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "425e8edd-c002-45a1-bb6f-ee3ac8812509" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 928.536418] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "425e8edd-c002-45a1-bb6f-ee3ac8812509" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.536703] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "425e8edd-c002-45a1-bb6f-ee3ac8812509-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 928.536952] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "425e8edd-c002-45a1-bb6f-ee3ac8812509-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.537169] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "425e8edd-c002-45a1-bb6f-ee3ac8812509-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.541975] env[62814]: INFO nova.compute.manager [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Terminating instance [ 928.601614] env[62814]: DEBUG nova.compute.manager [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Received event network-vif-plugged-f3e3a6f1-ae61-48eb-a135-02afe3f22971 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 928.601836] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] Acquiring lock "5c574786-64a4-4f07-a267-101ecaaa6938-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 928.602051] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] Lock "5c574786-64a4-4f07-a267-101ecaaa6938-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 928.602255] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] Lock "5c574786-64a4-4f07-a267-101ecaaa6938-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 928.602421] env[62814]: DEBUG nova.compute.manager [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] No waiting events found dispatching network-vif-plugged-f3e3a6f1-ae61-48eb-a135-02afe3f22971 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 928.602584] env[62814]: WARNING nova.compute.manager [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Received unexpected event network-vif-plugged-f3e3a6f1-ae61-48eb-a135-02afe3f22971 for instance with vm_state building and task_state spawning. [ 928.602740] env[62814]: DEBUG nova.compute.manager [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Received event network-changed-f3e3a6f1-ae61-48eb-a135-02afe3f22971 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 928.602891] env[62814]: DEBUG nova.compute.manager [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Refreshing instance network info cache due to event network-changed-f3e3a6f1-ae61-48eb-a135-02afe3f22971. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 928.603084] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] Acquiring lock "refresh_cache-5c574786-64a4-4f07-a267-101ecaaa6938" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.677579] env[62814]: DEBUG oslo_vmware.api [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4293997, 'name': PowerOnVM_Task, 'duration_secs': 0.613617} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.677579] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.679024] env[62814]: DEBUG nova.compute.manager [None req-8fe4d1ce-6e1d-4f6e-9647-f58c78e638a8 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 928.682444] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eceaabc7-5d55-4bb7-bd17-056fc115bce8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.785239] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad951c4-5800-4fa5-8020-4d08c7b3bd94 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.793326] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47f45ed-f66c-4144-a322-38f9e7e8fb97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.824634] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 928.829021] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685333c6-e249-4a13-bda8-31d91f058967 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.836393] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ed40a8-a92a-4901-91c9-ad189b2126dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.852679] env[62814]: DEBUG nova.compute.provider_tree [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 928.886740] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293998, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.912183] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Releasing lock "refresh_cache-5c574786-64a4-4f07-a267-101ecaaa6938" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 928.912288] env[62814]: DEBUG nova.compute.manager [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Instance network_info: |[{"id": "f3e3a6f1-ae61-48eb-a135-02afe3f22971", "address": "fa:16:3e:f9:d2:13", "network": {"id": "b16d5204-f140-44fb-b4c4-621c3f5a3423", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-309109771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5f0190bd5f24a909c9839b7fd1e870b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec46b14d-3310-4f2b-96c1-f53ee47d3759", "external-id": "nsx-vlan-transportzone-323", "segmentation_id": 323, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3e3a6f1-ae", "ovs_interfaceid": "f3e3a6f1-ae61-48eb-a135-02afe3f22971", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 928.912584] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] Acquired lock "refresh_cache-5c574786-64a4-4f07-a267-101ecaaa6938" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 928.912762] env[62814]: DEBUG nova.network.neutron [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Refreshing network info cache for port f3e3a6f1-ae61-48eb-a135-02afe3f22971 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 928.915411] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:d2:13', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ec46b14d-3310-4f2b-96c1-f53ee47d3759', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f3e3a6f1-ae61-48eb-a135-02afe3f22971', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 928.923034] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 928.924038] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 928.924322] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3bcec5be-0c8e-4dc9-88ac-8af2b65a74b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.946609] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 928.946609] env[62814]: value = "task-4293999" [ 928.946609] env[62814]: _type = "Task" [ 928.946609] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.950456] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 928.950741] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75e5e55c-4c48-4c28-adce-0aa676d71e1c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.957351] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293999, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.958872] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 928.958872] env[62814]: value = "task-4294000" [ 928.958872] env[62814]: _type = "Task" [ 928.958872] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.966932] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294000, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.047635] env[62814]: DEBUG nova.compute.manager [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 929.047635] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 929.048516] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae456607-6aa8-4da8-be3c-7e3f10ea1e8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.056783] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 929.057119] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef676ac5-68dd-4183-89fc-d2195d0bb52f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.065283] env[62814]: DEBUG oslo_vmware.api [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 929.065283] env[62814]: value = "task-4294001" [ 929.065283] env[62814]: _type = "Task" [ 929.065283] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.073851] env[62814]: DEBUG oslo_vmware.api [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4294001, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.091695] env[62814]: INFO nova.compute.manager [None req-6170fb6c-2a3c-424b-9c07-bfa204860ba4 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance to original state: 'active' [ 929.391570] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293998, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.392645] env[62814]: DEBUG nova.scheduler.client.report [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 101 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 929.392916] env[62814]: DEBUG nova.compute.provider_tree [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 101 to 102 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 929.393137] env[62814]: DEBUG nova.compute.provider_tree [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 142, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 929.459072] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4293999, 'name': CreateVM_Task, 'duration_secs': 0.352708} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.459072] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 929.459072] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.459072] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 929.459072] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 929.459511] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37a7b94d-52e6-463b-bc75-1d2df3444ef5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.471576] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 929.471576] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520aa23a-da0d-54a0-6a09-226f46884f13" [ 929.471576] env[62814]: _type = "Task" [ 929.471576] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.475404] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294000, 'name': PowerOffVM_Task, 'duration_secs': 0.140778} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.479528] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.480352] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 929.481506] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8aaa22-dd57-4ac1-94db-b468c7efb99d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.492151] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520aa23a-da0d-54a0-6a09-226f46884f13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.494572] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 929.494830] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7471257-3dec-4fdf-aeb6-5f97f0b7acc2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.521199] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 929.521421] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 929.521602] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Deleting the datastore file [datastore2] 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.521881] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78523c19-dc38-4089-a1d7-b324755b5014 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.528557] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 929.528557] env[62814]: value = "task-4294004" [ 929.528557] env[62814]: _type = "Task" [ 929.528557] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.536825] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294004, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.574481] env[62814]: DEBUG oslo_vmware.api [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4294001, 'name': PowerOffVM_Task, 'duration_secs': 0.220752} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.574742] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 929.574910] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 929.575373] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af5d531e-d3d8-4af9-b319-a99453b3fc6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.640129] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 929.640129] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 929.640129] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Deleting the datastore file [datastore2] 425e8edd-c002-45a1-bb6f-ee3ac8812509 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.640129] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7acc98ff-b863-4efa-8bd8-05a4cc3ab6da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.646602] env[62814]: DEBUG oslo_vmware.api [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for the task: (returnval){ [ 929.646602] env[62814]: value = "task-4294006" [ 929.646602] env[62814]: _type = "Task" [ 929.646602] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.656570] env[62814]: DEBUG oslo_vmware.api [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4294006, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.720841] env[62814]: DEBUG nova.network.neutron [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Updated VIF entry in instance network info cache for port f3e3a6f1-ae61-48eb-a135-02afe3f22971. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 929.721200] env[62814]: DEBUG nova.network.neutron [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Updating instance_info_cache with network_info: [{"id": "f3e3a6f1-ae61-48eb-a135-02afe3f22971", "address": "fa:16:3e:f9:d2:13", "network": {"id": "b16d5204-f140-44fb-b4c4-621c3f5a3423", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-309109771-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d5f0190bd5f24a909c9839b7fd1e870b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ec46b14d-3310-4f2b-96c1-f53ee47d3759", "external-id": "nsx-vlan-transportzone-323", "segmentation_id": 323, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3e3a6f1-ae", "ovs_interfaceid": "f3e3a6f1-ae61-48eb-a135-02afe3f22971", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.809868] env[62814]: INFO nova.compute.manager [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Rebuilding instance [ 929.852728] env[62814]: DEBUG nova.compute.manager [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 929.853798] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73151a3e-0573-4afc-871b-75773b3d4c34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.887894] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4293998, 'name': CreateSnapshot_Task, 'duration_secs': 1.057444} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.887894] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 929.888150] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab6220b-dcba-4b4b-af8e-a2cc383ed209 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.900881] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.410s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 929.901095] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 15.828s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 929.901491] env[62814]: DEBUG nova.objects.instance [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 929.924741] env[62814]: INFO nova.scheduler.client.report [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Deleted allocations for instance 1f40dc62-1a58-4cfb-8785-c37b68747f37 [ 929.986670] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520aa23a-da0d-54a0-6a09-226f46884f13, 'name': SearchDatastore_Task, 'duration_secs': 0.021949} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.987623] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 929.987893] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 929.988205] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.988384] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 929.988616] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 929.989222] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03209699-4802-4f84-806c-166fdfd429dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.998844] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 929.998971] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 930.000132] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-906533e8-5d46-4dc0-a8c8-0a3132f2453a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.009018] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 930.009018] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc8c14-4bec-8b30-a1d0-8c580162dbdb" [ 930.009018] env[62814]: _type = "Task" [ 930.009018] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.015524] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc8c14-4bec-8b30-a1d0-8c580162dbdb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.037776] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294004, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095682} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.038070] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.038288] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 930.038490] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 930.157019] env[62814]: DEBUG oslo_vmware.api [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Task: {'id': task-4294006, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151962} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.157432] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.157687] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 930.158878] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 930.158878] env[62814]: INFO nova.compute.manager [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Took 1.11 seconds to destroy the instance on the hypervisor. [ 930.158878] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 930.158878] env[62814]: DEBUG nova.compute.manager [-] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 930.158878] env[62814]: DEBUG nova.network.neutron [-] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 930.217269] env[62814]: INFO nova.compute.manager [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Unrescuing [ 930.217600] env[62814]: DEBUG oslo_concurrency.lockutils [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.217845] env[62814]: DEBUG oslo_concurrency.lockutils [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquired lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 930.218122] env[62814]: DEBUG nova.network.neutron [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.225864] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c83ead4-fe75-4df2-bf57-fe0b8d6db846 req-796986cc-fdf9-48ec-b9f3-242d838662ca service nova] Releasing lock "refresh_cache-5c574786-64a4-4f07-a267-101ecaaa6938" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 930.406580] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 930.412455] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d9c51c18-9cd8-4eb6-9acb-bebbd42abb54 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.426024] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 930.426024] env[62814]: value = "task-4294007" [ 930.426024] env[62814]: _type = "Task" [ 930.426024] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.432486] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294007, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.435506] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b9b3c3a7-b165-49ed-b4e4-a9adfe2edc38 tempest-ServersTestJSON-1841496840 tempest-ServersTestJSON-1841496840-project-member] Lock "1f40dc62-1a58-4cfb-8785-c37b68747f37" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.748s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.519481] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc8c14-4bec-8b30-a1d0-8c580162dbdb, 'name': SearchDatastore_Task, 'duration_secs': 0.020569} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.520345] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4dceb270-b8c2-4d7f-ac34-580271f3139b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.526837] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 930.526837] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f968f-9378-70bd-c7b3-377f41584784" [ 930.526837] env[62814]: _type = "Task" [ 930.526837] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.536361] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f968f-9378-70bd-c7b3-377f41584784, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.787609] env[62814]: DEBUG nova.compute.manager [req-93fc599a-a465-4642-8c8e-cf40947ea05c req-4f1d42ac-58ed-4338-a46c-5735f02fa0bd service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Received event network-vif-deleted-69314b3d-cb79-40f5-8562-afed90be617f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 930.787858] env[62814]: INFO nova.compute.manager [req-93fc599a-a465-4642-8c8e-cf40947ea05c req-4f1d42ac-58ed-4338-a46c-5735f02fa0bd service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Neutron deleted interface 69314b3d-cb79-40f5-8562-afed90be617f; detaching it from the instance and deleting it from the info cache [ 930.788065] env[62814]: DEBUG nova.network.neutron [req-93fc599a-a465-4642-8c8e-cf40947ea05c req-4f1d42ac-58ed-4338-a46c-5735f02fa0bd service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.868314] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.869112] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e8e48d36-e5bc-406f-bd16-6bd9154d3760 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.877053] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 930.877293] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.877941] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 930.877941] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.877941] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.879647] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 930.879647] env[62814]: value = "task-4294008" [ 930.879647] env[62814]: _type = "Task" [ 930.879647] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.880078] env[62814]: INFO nova.compute.manager [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Terminating instance [ 930.894179] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294008, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.911208] env[62814]: DEBUG nova.network.neutron [-] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.918414] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e72fbae6-6c2e-424e-a7d3-6428e76c7ee5 tempest-ServersAdmin275Test-1788413043 tempest-ServersAdmin275Test-1788413043-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 930.919241] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.083s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 930.920810] env[62814]: INFO nova.compute.claims [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.940929] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294007, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.038351] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f968f-9378-70bd-c7b3-377f41584784, 'name': SearchDatastore_Task, 'duration_secs': 0.01194} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.038674] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.038986] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 5c574786-64a4-4f07-a267-101ecaaa6938/5c574786-64a4-4f07-a267-101ecaaa6938.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 931.039687] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eb26102e-4166-4ca8-9548-3f8408447327 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.051571] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 931.051571] env[62814]: value = "task-4294009" [ 931.051571] env[62814]: _type = "Task" [ 931.051571] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.061873] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294009, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.074400] env[62814]: DEBUG nova.network.neutron [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updating instance_info_cache with network_info: [{"id": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "address": "fa:16:3e:aa:55:02", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b486fb-bb", "ovs_interfaceid": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.084269] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 931.084692] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 931.086094] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 931.086094] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 931.086435] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 931.086636] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 931.086868] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 931.087072] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 931.087284] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 931.088073] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 931.088073] env[62814]: DEBUG nova.virt.hardware [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 931.088583] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d68cb675-59a7-4b3c-922b-1892e3773bbd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.099153] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6afbdfa6-1987-47d4-86fa-d89fd93d6b8d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.115128] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 931.121371] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 931.121371] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 931.121606] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ba25d04-845b-4a4b-9576-c7194476d271 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.138944] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 931.138944] env[62814]: value = "task-4294010" [ 931.138944] env[62814]: _type = "Task" [ 931.138944] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.147583] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294010, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.292073] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2469f117-70bd-464d-8def-bf5eecca8461 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.300539] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fc7d010-77b8-4264-a631-8364eea96437 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.338652] env[62814]: DEBUG nova.compute.manager [req-93fc599a-a465-4642-8c8e-cf40947ea05c req-4f1d42ac-58ed-4338-a46c-5735f02fa0bd service nova] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Detach interface failed, port_id=69314b3d-cb79-40f5-8562-afed90be617f, reason: Instance 425e8edd-c002-45a1-bb6f-ee3ac8812509 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 931.390987] env[62814]: DEBUG nova.compute.manager [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 931.391146] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 931.392358] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294008, 'name': PowerOffVM_Task, 'duration_secs': 0.305551} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.392358] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f51e69-63e3-40c5-a613-adac25b68bb4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.395716] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 931.395957] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 931.397338] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb0c2a4-8c32-4002-af09-86dde9100820 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.406385] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.408541] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-869e34a0-28cf-4d95-b07a-d5b8e2cac2ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.410630] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 931.411332] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4ba8b6f-6224-4ece-9a8e-a56b6f91ca9c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.413370] env[62814]: INFO nova.compute.manager [-] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Took 1.25 seconds to deallocate network for instance. [ 931.427031] env[62814]: DEBUG oslo_vmware.api [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 931.427031] env[62814]: value = "task-4294011" [ 931.427031] env[62814]: _type = "Task" [ 931.427031] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.444085] env[62814]: DEBUG oslo_vmware.api [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4294011, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.449142] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294007, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.485156] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 931.485447] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 931.486102] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleting the datastore file [datastore2] 98ba2786-023f-4d36-bdd0-e38bdd23d73c {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.486102] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-da28501f-ce13-4c0b-83c6-1b6c644e2dbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.501471] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 931.501471] env[62814]: value = "task-4294013" [ 931.501471] env[62814]: _type = "Task" [ 931.501471] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.511580] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294013, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.563145] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294009, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.578349] env[62814]: DEBUG oslo_concurrency.lockutils [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Releasing lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.578605] env[62814]: DEBUG nova.objects.instance [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lazy-loading 'flavor' on Instance uuid 3e4479a0-8edd-4b37-8cc9-2c91275b88ee {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.645728] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 931.646105] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845740', 'volume_id': '4bf78821-2028-4edc-a038-062cb9d09c71', 'name': 'volume-4bf78821-2028-4edc-a038-062cb9d09c71', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1e259ec6-d31a-453b-87e0-baa446665d56', 'attached_at': '', 'detached_at': '', 'volume_id': '4bf78821-2028-4edc-a038-062cb9d09c71', 'serial': '4bf78821-2028-4edc-a038-062cb9d09c71'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 931.647201] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf4247f-383e-4e6b-bb53-5fc6bfad4298 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.653231] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294010, 'name': CreateVM_Task, 'duration_secs': 0.386024} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.653853] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 931.654362] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.654802] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 931.654994] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 931.655323] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-227a3a59-870c-471c-acd7-db889eebfef0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.670771] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98b8818-048e-48c3-83ad-d78edb829880 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.674913] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 931.674913] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e5de98-601e-adaa-2a4a-b8b044d18593" [ 931.674913] env[62814]: _type = "Task" [ 931.674913] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.700272] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] volume-4bf78821-2028-4edc-a038-062cb9d09c71/volume-4bf78821-2028-4edc-a038-062cb9d09c71.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 931.702362] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6090d4ad-c6e2-47fc-b8a8-0a3be44c7f98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.720342] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e5de98-601e-adaa-2a4a-b8b044d18593, 'name': SearchDatastore_Task, 'duration_secs': 0.009652} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.721206] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 931.721518] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.721795] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.721981] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 931.722217] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.722520] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-292b4b7d-792a-4d4c-a41a-1560773103d6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.727545] env[62814]: DEBUG oslo_vmware.api [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 931.727545] env[62814]: value = "task-4294014" [ 931.727545] env[62814]: _type = "Task" [ 931.727545] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.753604] env[62814]: DEBUG oslo_vmware.api [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294014, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.755482] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.755724] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 931.756759] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c74d23b-0c08-4e82-a6c1-06e4c917d378 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.763547] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 931.763547] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527db00c-6d4d-9358-cceb-8bc578308355" [ 931.763547] env[62814]: _type = "Task" [ 931.763547] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.774146] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527db00c-6d4d-9358-cceb-8bc578308355, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.925528] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 931.953732] env[62814]: DEBUG oslo_vmware.api [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4294011, 'name': PowerOffVM_Task, 'duration_secs': 0.281505} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.953949] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294007, 'name': CloneVM_Task} progress is 95%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.954250] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 931.954660] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 931.955018] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3030a13f-caa5-40a2-9fa6-802fac4fa3a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.014780] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294013, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.287801} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.015242] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.016928] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 932.016928] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 932.027532] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 932.027641] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 932.027918] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Deleting the datastore file [datastore2] 9cd89da4-d190-4bfc-81e1-da15b98d54ba {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 932.028360] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5528ba8-87e3-4a0a-b12a-e9936b3c0cf9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.041189] env[62814]: DEBUG oslo_vmware.api [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 932.041189] env[62814]: value = "task-4294016" [ 932.041189] env[62814]: _type = "Task" [ 932.041189] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.054884] env[62814]: DEBUG oslo_vmware.api [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4294016, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.063933] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294009, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587818} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.066091] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 5c574786-64a4-4f07-a267-101ecaaa6938/5c574786-64a4-4f07-a267-101ecaaa6938.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.066225] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.067083] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f0e4155-7fc7-4638-8873-355627e07128 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.072766] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 932.072766] env[62814]: value = "task-4294017" [ 932.072766] env[62814]: _type = "Task" [ 932.072766] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.088877] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294017, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.089646] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a00979f-d135-41a4-bf11-8087818b9d94 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.120862] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.120862] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47751d16-7c2c-4630-9049-593ed88a50ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.127851] env[62814]: DEBUG oslo_vmware.api [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 932.127851] env[62814]: value = "task-4294018" [ 932.127851] env[62814]: _type = "Task" [ 932.127851] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.137923] env[62814]: DEBUG oslo_vmware.api [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.240692] env[62814]: DEBUG oslo_vmware.api [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294014, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.278647] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527db00c-6d4d-9358-cceb-8bc578308355, 'name': SearchDatastore_Task, 'duration_secs': 0.01812} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.279576] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f025f45-0ae5-4851-ace8-8f94cde3e2d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.287489] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 932.287489] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5203f4f7-22ed-1390-e57d-81f359e68575" [ 932.287489] env[62814]: _type = "Task" [ 932.287489] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.295903] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5203f4f7-22ed-1390-e57d-81f359e68575, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.435158] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294007, 'name': CloneVM_Task, 'duration_secs': 1.725732} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.438899] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Created linked-clone VM from snapshot [ 932.440065] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c580a59-9df0-4814-bd60-d8b928960654 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.447446] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Uploading image af88e617-86b6-4046-9983-8f685578152f {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 932.473613] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 932.473613] env[62814]: value = "vm-845743" [ 932.473613] env[62814]: _type = "VirtualMachine" [ 932.473613] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 932.473915] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8129a2f3-6ec6-4dc7-b53e-91ba6095e53b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.477696] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9643d1-8e2f-4651-81ac-4f90c2e5d695 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.481463] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lease: (returnval){ [ 932.481463] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522d89f7-fee3-ae8e-068e-3abeb5214ec9" [ 932.481463] env[62814]: _type = "HttpNfcLease" [ 932.481463] env[62814]: } obtained for exporting VM: (result){ [ 932.481463] env[62814]: value = "vm-845743" [ 932.481463] env[62814]: _type = "VirtualMachine" [ 932.481463] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 932.481743] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the lease: (returnval){ [ 932.481743] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522d89f7-fee3-ae8e-068e-3abeb5214ec9" [ 932.481743] env[62814]: _type = "HttpNfcLease" [ 932.481743] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 932.488241] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f251d68e-bf93-4f89-a4dc-355c1be7f896 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.492745] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 932.492745] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522d89f7-fee3-ae8e-068e-3abeb5214ec9" [ 932.492745] env[62814]: _type = "HttpNfcLease" [ 932.492745] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 932.526574] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bb9b262-763f-45f6-b0ec-02de348a7416 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.541862] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1576a6d9-05c9-492b-a580-e8458db7e784 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.567658] env[62814]: DEBUG nova.compute.provider_tree [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.569734] env[62814]: DEBUG oslo_vmware.api [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4294016, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.328541} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.569734] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.569734] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 932.569734] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 932.570093] env[62814]: INFO nova.compute.manager [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Took 1.18 seconds to destroy the instance on the hypervisor. [ 932.570093] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 932.570646] env[62814]: DEBUG nova.compute.manager [-] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 932.570646] env[62814]: DEBUG nova.network.neutron [-] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 932.590378] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294017, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072455} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.590378] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 932.590378] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34dc3e70-7214-407f-8dc8-f36546846a6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.616336] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 5c574786-64a4-4f07-a267-101ecaaa6938/5c574786-64a4-4f07-a267-101ecaaa6938.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.616632] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7217e90-7d87-4bec-af8b-62681c4049d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.642717] env[62814]: DEBUG oslo_vmware.api [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294018, 'name': PowerOffVM_Task, 'duration_secs': 0.478867} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.644093] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 932.650043] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Reconfiguring VM instance instance-0000002b to detach disk 2002 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 932.650430] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 932.650430] env[62814]: value = "task-4294020" [ 932.650430] env[62814]: _type = "Task" [ 932.650430] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.651067] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cfd0e635-ae7b-475a-9364-18f6c5016526 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.677045] env[62814]: DEBUG oslo_vmware.api [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 932.677045] env[62814]: value = "task-4294021" [ 932.677045] env[62814]: _type = "Task" [ 932.677045] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.686146] env[62814]: DEBUG oslo_vmware.api [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294021, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.745313] env[62814]: DEBUG oslo_vmware.api [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294014, 'name': ReconfigVM_Task, 'duration_secs': 0.630061} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.745742] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Reconfigured VM instance instance-0000003c to attach disk [datastore1] volume-4bf78821-2028-4edc-a038-062cb9d09c71/volume-4bf78821-2028-4edc-a038-062cb9d09c71.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 932.751066] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82c75024-5904-400b-9184-c76b8cf1965a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.767616] env[62814]: DEBUG oslo_vmware.api [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 932.767616] env[62814]: value = "task-4294022" [ 932.767616] env[62814]: _type = "Task" [ 932.767616] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.776313] env[62814]: DEBUG oslo_vmware.api [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294022, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.797941] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5203f4f7-22ed-1390-e57d-81f359e68575, 'name': SearchDatastore_Task, 'duration_secs': 0.01306} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.798248] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 932.798516] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a/0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 932.798788] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e97dcd78-12f5-4f1b-870d-4f80868497e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.805676] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 932.805676] env[62814]: value = "task-4294023" [ 932.805676] env[62814]: _type = "Task" [ 932.805676] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.813333] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294023, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.991417] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 932.991417] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522d89f7-fee3-ae8e-068e-3abeb5214ec9" [ 932.991417] env[62814]: _type = "HttpNfcLease" [ 932.991417] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 932.991768] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 932.991768] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522d89f7-fee3-ae8e-068e-3abeb5214ec9" [ 932.991768] env[62814]: _type = "HttpNfcLease" [ 932.991768] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 932.993963] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160cf012-9bc4-4066-8124-71d8fbb6bdd6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.000860] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52081238-322c-735e-b8ae-0e13acb20987/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 933.001081] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52081238-322c-735e-b8ae-0e13acb20987/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 933.099077] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-bed358d4-8b99-421d-a913-8b2c1712c481 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.120170] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 933.120516] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.121011] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 933.121011] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.121011] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 933.121162] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 933.121584] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 933.121584] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 933.121745] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 933.121899] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 933.122124] env[62814]: DEBUG nova.virt.hardware [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 933.123071] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-661ba294-afe8-4a70-a04a-a262fd75dd9d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.129926] env[62814]: DEBUG nova.scheduler.client.report [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 102 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 933.131340] env[62814]: DEBUG nova.compute.provider_tree [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 102 to 103 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 933.131340] env[62814]: DEBUG nova.compute.provider_tree [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 933.142077] env[62814]: DEBUG nova.compute.manager [req-41e931be-a70c-4a89-a6c1-3dc6eaa6f1ac req-c535b94e-db91-4f4a-a61b-7ad8e9d27979 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Received event network-vif-deleted-bc38976c-8838-47e6-b4f8-cd013e62b15b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 933.142317] env[62814]: INFO nova.compute.manager [req-41e931be-a70c-4a89-a6c1-3dc6eaa6f1ac req-c535b94e-db91-4f4a-a61b-7ad8e9d27979 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Neutron deleted interface bc38976c-8838-47e6-b4f8-cd013e62b15b; detaching it from the instance and deleting it from the info cache [ 933.142592] env[62814]: DEBUG nova.network.neutron [req-41e931be-a70c-4a89-a6c1-3dc6eaa6f1ac req-c535b94e-db91-4f4a-a61b-7ad8e9d27979 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.149700] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94486c4b-be76-4cb6-8162-a0e5aecacfee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.168544] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:37:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '742fe6f0-d787-4ca8-be9a-edabfb0646a0', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 933.176192] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 933.185311] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 933.186080] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa2c951e-3045-40f2-a162-faf858628eea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.209168] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294020, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.212844] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 933.212844] env[62814]: value = "task-4294024" [ 933.212844] env[62814]: _type = "Task" [ 933.212844] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.213127] env[62814]: DEBUG oslo_vmware.api [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294021, 'name': ReconfigVM_Task, 'duration_secs': 0.334458} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.213653] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Reconfigured VM instance instance-0000002b to detach disk 2002 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 933.213653] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 933.217557] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-48ad06b8-c90f-4818-a064-8a67a277e32c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.224325] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294024, 'name': CreateVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.225493] env[62814]: DEBUG oslo_vmware.api [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 933.225493] env[62814]: value = "task-4294025" [ 933.225493] env[62814]: _type = "Task" [ 933.225493] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.232783] env[62814]: DEBUG oslo_vmware.api [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294025, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.281042] env[62814]: DEBUG oslo_vmware.api [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.316309] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294023, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.637018] env[62814]: DEBUG nova.network.neutron [-] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.640559] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.721s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 933.641622] env[62814]: DEBUG nova.compute.manager [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 933.644801] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.350s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 933.649026] env[62814]: INFO nova.compute.claims [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.652341] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b76817f5-4611-4acb-ba47-49aaa24ecc02 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.665424] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d4ac6f-65cc-4127-bc2c-2a98f8130410 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.690476] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294020, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.718942] env[62814]: DEBUG nova.compute.manager [req-41e931be-a70c-4a89-a6c1-3dc6eaa6f1ac req-c535b94e-db91-4f4a-a61b-7ad8e9d27979 service nova] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Detach interface failed, port_id=bc38976c-8838-47e6-b4f8-cd013e62b15b, reason: Instance 9cd89da4-d190-4bfc-81e1-da15b98d54ba could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 933.730200] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294024, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.739296] env[62814]: DEBUG oslo_vmware.api [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294025, 'name': PowerOnVM_Task, 'duration_secs': 0.463421} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.739584] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 933.739817] env[62814]: DEBUG nova.compute.manager [None req-45e4233d-8794-451a-8d54-65824f6443f9 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 933.741667] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f9b9306-c4e0-45b3-88b2-45e77c1c62f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.785522] env[62814]: DEBUG oslo_vmware.api [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294022, 'name': ReconfigVM_Task, 'duration_secs': 0.696841} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.786096] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845740', 'volume_id': '4bf78821-2028-4edc-a038-062cb9d09c71', 'name': 'volume-4bf78821-2028-4edc-a038-062cb9d09c71', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1e259ec6-d31a-453b-87e0-baa446665d56', 'attached_at': '', 'detached_at': '', 'volume_id': '4bf78821-2028-4edc-a038-062cb9d09c71', 'serial': '4bf78821-2028-4edc-a038-062cb9d09c71'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 933.817505] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294023, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.142618] env[62814]: INFO nova.compute.manager [-] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Took 1.57 seconds to deallocate network for instance. [ 934.154470] env[62814]: DEBUG nova.compute.utils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 934.158482] env[62814]: DEBUG nova.compute.manager [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 934.158840] env[62814]: DEBUG nova.network.neutron [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 934.196851] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294020, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.228444] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294024, 'name': CreateVM_Task, 'duration_secs': 0.729295} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.228444] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 934.229159] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.229401] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 934.229978] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 934.230093] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc9ee543-a556-4e7e-8660-66774168f612 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.234811] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 934.234811] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a8bcf0-621a-2674-d74c-0168905e4a1a" [ 934.234811] env[62814]: _type = "Task" [ 934.234811] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.242712] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a8bcf0-621a-2674-d74c-0168905e4a1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.258163] env[62814]: DEBUG nova.policy [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd26dbcbc80942b1a73c915fa9b31fda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a7d7cebe11d549aba28cc384eabcd863', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 934.319244] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294023, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.012921} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.319674] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a/0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.319970] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.320382] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e93ab180-3272-4d5c-9586-e9490bef288e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.327520] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 934.327520] env[62814]: value = "task-4294026" [ 934.327520] env[62814]: _type = "Task" [ 934.327520] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.341278] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294026, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.652349] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 934.659305] env[62814]: DEBUG nova.compute.manager [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 934.697896] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294020, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.751368] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a8bcf0-621a-2674-d74c-0168905e4a1a, 'name': SearchDatastore_Task, 'duration_secs': 0.02087} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.751975] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 934.752467] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.752798] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.753054] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 934.753346] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.753729] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de2882a8-c955-449a-acd0-3b57324f7266 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.766798] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.766798] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 934.769824] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0bdd22d-12ff-480e-881a-51d10f03fc49 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.773160] env[62814]: DEBUG nova.network.neutron [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Successfully created port: 20070e44-989b-4d3c-aab6-3de905aad547 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.781537] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 934.781537] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d3ed19-7b0b-6979-84bb-90f2ad8bb8e9" [ 934.781537] env[62814]: _type = "Task" [ 934.781537] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.792481] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d3ed19-7b0b-6979-84bb-90f2ad8bb8e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.837390] env[62814]: DEBUG nova.objects.instance [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lazy-loading 'flavor' on Instance uuid 1e259ec6-d31a-453b-87e0-baa446665d56 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.839553] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294026, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.195918] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294020, 'name': ReconfigVM_Task, 'duration_secs': 2.402027} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.199536] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 5c574786-64a4-4f07-a267-101ecaaa6938/5c574786-64a4-4f07-a267-101ecaaa6938.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.201132] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c26ae9b3-0aaa-427c-bb3b-d835bab375dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.211436] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 935.211436] env[62814]: value = "task-4294027" [ 935.211436] env[62814]: _type = "Task" [ 935.211436] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.222968] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294027, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.301950] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d3ed19-7b0b-6979-84bb-90f2ad8bb8e9, 'name': SearchDatastore_Task, 'duration_secs': 0.015177} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.303727] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e16e2555-1bfd-4db3-949e-2fd5edc5be98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.310287] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 935.310287] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5203eef2-871b-0c46-33e2-af24413c1aab" [ 935.310287] env[62814]: _type = "Task" [ 935.310287] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.319953] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5203eef2-871b-0c46-33e2-af24413c1aab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.321860] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613253b2-5728-4233-b724-c4aecabb8e19 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.333330] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0894c4a0-dbec-4f85-a0b0-1dbd3e48a1ce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.345019] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294026, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.754849} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.372728] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d8195956-cc3c-4728-a98d-c12669997d4a tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.420s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 935.374245] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 935.375566] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f6143aa-8614-42ff-81f5-613f012f336c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.381574] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7824cf-d9c3-410a-907e-c52f41925a70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.409227] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a/0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.413435] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6890942-a732-4b60-9e54-de4579224739 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.429417] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5ec63ab-03b7-49fc-9310-573adc258a81 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.444832] env[62814]: DEBUG nova.compute.provider_tree [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.447817] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 935.447817] env[62814]: value = "task-4294028" [ 935.447817] env[62814]: _type = "Task" [ 935.447817] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.674135] env[62814]: DEBUG nova.compute.manager [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 935.705661] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 935.708046] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.708046] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 935.708046] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.708046] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 935.708046] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 935.708046] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 935.708046] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 935.708046] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 935.708046] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 935.708046] env[62814]: DEBUG nova.virt.hardware [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 935.708540] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2df6752-9a43-4b36-8fd8-78029dc42ebd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.719740] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd5f5bb-70b2-4ab6-9b60-e731876f5f67 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.736094] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294027, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.823229] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5203eef2-871b-0c46-33e2-af24413c1aab, 'name': SearchDatastore_Task, 'duration_secs': 0.020373} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.823598] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 935.823924] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98ba2786-023f-4d36-bdd0-e38bdd23d73c/98ba2786-023f-4d36-bdd0-e38bdd23d73c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 935.824338] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-944a6040-f797-4be8-8639-e13bb6c52d51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.834460] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 935.834460] env[62814]: value = "task-4294029" [ 935.834460] env[62814]: _type = "Task" [ 935.834460] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.845829] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294029, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.857371] env[62814]: DEBUG nova.compute.manager [req-e925f112-c27b-4970-9d71-ce0b1f5f41b1 req-e57f9b47-3f7e-4900-8c8b-a71d38d93381 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Received event network-changed-95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 935.857575] env[62814]: DEBUG nova.compute.manager [req-e925f112-c27b-4970-9d71-ce0b1f5f41b1 req-e57f9b47-3f7e-4900-8c8b-a71d38d93381 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Refreshing instance network info cache due to event network-changed-95b486fb-bbe0-4666-a926-01bcfa37eaad. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 935.857873] env[62814]: DEBUG oslo_concurrency.lockutils [req-e925f112-c27b-4970-9d71-ce0b1f5f41b1 req-e57f9b47-3f7e-4900-8c8b-a71d38d93381 service nova] Acquiring lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.858048] env[62814]: DEBUG oslo_concurrency.lockutils [req-e925f112-c27b-4970-9d71-ce0b1f5f41b1 req-e57f9b47-3f7e-4900-8c8b-a71d38d93381 service nova] Acquired lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 935.858322] env[62814]: DEBUG nova.network.neutron [req-e925f112-c27b-4970-9d71-ce0b1f5f41b1 req-e57f9b47-3f7e-4900-8c8b-a71d38d93381 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Refreshing network info cache for port 95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 935.950236] env[62814]: DEBUG nova.scheduler.client.report [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 935.973530] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294028, 'name': ReconfigVM_Task, 'duration_secs': 0.455616} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.974037] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a/0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.974971] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96042e6d-dda2-47b0-9b86-fb2d1691e4e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.982933] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 935.982933] env[62814]: value = "task-4294030" [ 935.982933] env[62814]: _type = "Task" [ 935.982933] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.996044] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294030, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.222713] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294027, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.344082] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294029, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.461106] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.814s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.461106] env[62814]: DEBUG nova.compute.manager [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 936.462247] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.878s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 936.462620] env[62814]: DEBUG nova.objects.instance [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 936.471620] env[62814]: DEBUG nova.compute.manager [req-d2cf4b27-399e-41d3-8f3f-204bbdceb725 req-3dbcbe83-87be-41a3-bc60-7c3ef1d788e6 service nova] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Received event network-vif-plugged-20070e44-989b-4d3c-aab6-3de905aad547 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 936.476121] env[62814]: DEBUG oslo_concurrency.lockutils [req-d2cf4b27-399e-41d3-8f3f-204bbdceb725 req-3dbcbe83-87be-41a3-bc60-7c3ef1d788e6 service nova] Acquiring lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 936.476121] env[62814]: DEBUG oslo_concurrency.lockutils [req-d2cf4b27-399e-41d3-8f3f-204bbdceb725 req-3dbcbe83-87be-41a3-bc60-7c3ef1d788e6 service nova] Lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 936.476121] env[62814]: DEBUG oslo_concurrency.lockutils [req-d2cf4b27-399e-41d3-8f3f-204bbdceb725 req-3dbcbe83-87be-41a3-bc60-7c3ef1d788e6 service nova] Lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 936.476121] env[62814]: DEBUG nova.compute.manager [req-d2cf4b27-399e-41d3-8f3f-204bbdceb725 req-3dbcbe83-87be-41a3-bc60-7c3ef1d788e6 service nova] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] No waiting events found dispatching network-vif-plugged-20070e44-989b-4d3c-aab6-3de905aad547 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 936.480067] env[62814]: WARNING nova.compute.manager [req-d2cf4b27-399e-41d3-8f3f-204bbdceb725 req-3dbcbe83-87be-41a3-bc60-7c3ef1d788e6 service nova] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Received unexpected event network-vif-plugged-20070e44-989b-4d3c-aab6-3de905aad547 for instance with vm_state building and task_state spawning. [ 936.493796] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294030, 'name': Rename_Task, 'duration_secs': 0.173421} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.496815] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.496815] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a6404a0-ecc8-41f9-ae15-d2cdf724e373 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.500422] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 936.500422] env[62814]: value = "task-4294031" [ 936.500422] env[62814]: _type = "Task" [ 936.500422] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.509290] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294031, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.611799] env[62814]: DEBUG nova.network.neutron [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Successfully updated port: 20070e44-989b-4d3c-aab6-3de905aad547 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.723525] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294027, 'name': Rename_Task, 'duration_secs': 1.210003} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.725170] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 936.725170] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1354c2a3-8834-4fa0-92b9-91d1745a6034 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.731472] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 936.731472] env[62814]: value = "task-4294032" [ 936.731472] env[62814]: _type = "Task" [ 936.731472] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.746132] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294032, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.847712] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294029, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.847712] env[62814]: DEBUG nova.network.neutron [req-e925f112-c27b-4970-9d71-ce0b1f5f41b1 req-e57f9b47-3f7e-4900-8c8b-a71d38d93381 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updated VIF entry in instance network info cache for port 95b486fb-bbe0-4666-a926-01bcfa37eaad. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 936.847712] env[62814]: DEBUG nova.network.neutron [req-e925f112-c27b-4970-9d71-ce0b1f5f41b1 req-e57f9b47-3f7e-4900-8c8b-a71d38d93381 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updating instance_info_cache with network_info: [{"id": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "address": "fa:16:3e:aa:55:02", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b486fb-bb", "ovs_interfaceid": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.973715] env[62814]: DEBUG nova.compute.utils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 936.975568] env[62814]: DEBUG nova.compute.manager [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 936.975839] env[62814]: DEBUG nova.network.neutron [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 937.013436] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294031, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.101039] env[62814]: DEBUG nova.policy [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd26dbcbc80942b1a73c915fa9b31fda', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a7d7cebe11d549aba28cc384eabcd863', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 937.114353] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "refresh_cache-f54b8bb9-69b7-4bb4-a82c-9f796050e719" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.114353] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquired lock "refresh_cache-f54b8bb9-69b7-4bb4-a82c-9f796050e719" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 937.114353] env[62814]: DEBUG nova.network.neutron [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 937.242875] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294032, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.346331] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294029, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.055809} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.346630] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 98ba2786-023f-4d36-bdd0-e38bdd23d73c/98ba2786-023f-4d36-bdd0-e38bdd23d73c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 937.346847] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 937.347124] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5dbbcda-d59b-4582-86a0-815fe78a4348 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.351225] env[62814]: DEBUG oslo_concurrency.lockutils [req-e925f112-c27b-4970-9d71-ce0b1f5f41b1 req-e57f9b47-3f7e-4900-8c8b-a71d38d93381 service nova] Releasing lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 937.353998] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 937.353998] env[62814]: value = "task-4294033" [ 937.353998] env[62814]: _type = "Task" [ 937.353998] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.363720] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294033, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.484399] env[62814]: DEBUG nova.compute.manager [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 937.492742] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f27d165d-9a8e-433d-88a1-25def6a3f863 tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.030s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 937.493660] env[62814]: DEBUG oslo_concurrency.lockutils [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.493s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.496018] env[62814]: DEBUG nova.objects.instance [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lazy-loading 'resources' on Instance uuid cd89e4d5-c08e-42c9-aea8-84c20714e196 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.517814] env[62814]: DEBUG oslo_vmware.api [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294031, 'name': PowerOnVM_Task, 'duration_secs': 0.780514} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.518112] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.518324] env[62814]: DEBUG nova.compute.manager [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 937.519139] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca685fbd-f6cf-4edc-8c78-013d9a59fa7e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.533596] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "40e2d845-0211-4c84-aef7-94014f999e1d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 937.533845] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "40e2d845-0211-4c84-aef7-94014f999e1d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 937.581376] env[62814]: DEBUG nova.network.neutron [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Successfully created port: 5992ceda-be00-450a-a26e-10cab2e2a02b {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 937.648831] env[62814]: DEBUG nova.network.neutron [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.746126] env[62814]: DEBUG oslo_vmware.api [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294032, 'name': PowerOnVM_Task, 'duration_secs': 0.681291} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.746483] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.746784] env[62814]: INFO nova.compute.manager [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Took 11.36 seconds to spawn the instance on the hypervisor. [ 937.746834] env[62814]: DEBUG nova.compute.manager [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 937.747596] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0501defc-0db7-4868-9aa3-b3684e5da0b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.803538] env[62814]: DEBUG nova.network.neutron [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Updating instance_info_cache with network_info: [{"id": "20070e44-989b-4d3c-aab6-3de905aad547", "address": "fa:16:3e:bb:93:a1", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20070e44-98", "ovs_interfaceid": "20070e44-989b-4d3c-aab6-3de905aad547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.863869] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294033, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113866} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.864171] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.864983] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1433bb44-f9b5-476f-901c-d8dff2581eba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.897494] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 98ba2786-023f-4d36-bdd0-e38bdd23d73c/98ba2786-023f-4d36-bdd0-e38bdd23d73c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.898180] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f47086e2-6b0b-4ea3-923c-855ebc209f00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.914393] env[62814]: DEBUG nova.compute.manager [req-c009f7f9-dba0-49c2-a956-ec8bea241d24 req-578b5100-f612-4b98-a93d-ea0ab796ad2c service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Received event network-changed-95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 937.914657] env[62814]: DEBUG nova.compute.manager [req-c009f7f9-dba0-49c2-a956-ec8bea241d24 req-578b5100-f612-4b98-a93d-ea0ab796ad2c service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Refreshing instance network info cache due to event network-changed-95b486fb-bbe0-4666-a926-01bcfa37eaad. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 937.914794] env[62814]: DEBUG oslo_concurrency.lockutils [req-c009f7f9-dba0-49c2-a956-ec8bea241d24 req-578b5100-f612-4b98-a93d-ea0ab796ad2c service nova] Acquiring lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.914933] env[62814]: DEBUG oslo_concurrency.lockutils [req-c009f7f9-dba0-49c2-a956-ec8bea241d24 req-578b5100-f612-4b98-a93d-ea0ab796ad2c service nova] Acquired lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 937.915101] env[62814]: DEBUG nova.network.neutron [req-c009f7f9-dba0-49c2-a956-ec8bea241d24 req-578b5100-f612-4b98-a93d-ea0ab796ad2c service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Refreshing network info cache for port 95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 937.921617] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 937.921617] env[62814]: value = "task-4294034" [ 937.921617] env[62814]: _type = "Task" [ 937.921617] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.930588] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294034, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.038834] env[62814]: DEBUG nova.compute.manager [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 938.052499] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.272786] env[62814]: INFO nova.compute.manager [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Took 35.93 seconds to build instance. [ 938.306342] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Releasing lock "refresh_cache-f54b8bb9-69b7-4bb4-a82c-9f796050e719" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 938.306671] env[62814]: DEBUG nova.compute.manager [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Instance network_info: |[{"id": "20070e44-989b-4d3c-aab6-3de905aad547", "address": "fa:16:3e:bb:93:a1", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20070e44-98", "ovs_interfaceid": "20070e44-989b-4d3c-aab6-3de905aad547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 938.310228] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bb:93:a1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '20070e44-989b-4d3c-aab6-3de905aad547', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.317921] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Creating folder: Project (a7d7cebe11d549aba28cc384eabcd863). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 938.321029] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b89a805c-9ff3-48db-b2a2-48aa145d5ca9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.337645] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Created folder: Project (a7d7cebe11d549aba28cc384eabcd863) in parent group-v845547. [ 938.337645] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Creating folder: Instances. Parent ref: group-v845746. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 938.337645] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c72a092b-4458-4534-9410-6593fcbed672 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.346689] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Created folder: Instances in parent group-v845746. [ 938.346972] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 938.347316] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 938.347475] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f01dbe1a-3dda-484f-b97f-96779bab6c0a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.377406] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.377406] env[62814]: value = "task-4294037" [ 938.377406] env[62814]: _type = "Task" [ 938.377406] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.389328] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294037, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.438266] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294034, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.500251] env[62814]: DEBUG nova.compute.manager [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 938.538039] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 938.538039] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.538039] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 938.538039] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.538039] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 938.538039] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 938.538039] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 938.538039] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 938.538039] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 938.538039] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 938.538404] env[62814]: DEBUG nova.virt.hardware [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 938.540332] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5765f57-fc9c-40fd-bd72-5c9b3889e21d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.559588] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4bf58e-68fe-40ab-8d88-3af1a1b4ffac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.584288] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 938.664103] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9128aa1-d601-487d-b237-c1990a4a1bcb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.673194] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5dd1bd5-c8b5-4a41-8e87-e7fbc1b9edcb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.709830] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe92e72-1181-46c8-93f0-dcdc9bb935dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.717487] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ecadcfe-718f-4259-97a1-685c31183a61 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.733909] env[62814]: DEBUG nova.compute.provider_tree [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.775578] env[62814]: DEBUG oslo_concurrency.lockutils [None req-87854a8b-a487-4b48-91a2-9d781df5a332 tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "5c574786-64a4-4f07-a267-101ecaaa6938" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.443s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 938.819920] env[62814]: DEBUG nova.network.neutron [req-c009f7f9-dba0-49c2-a956-ec8bea241d24 req-578b5100-f612-4b98-a93d-ea0ab796ad2c service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updated VIF entry in instance network info cache for port 95b486fb-bbe0-4666-a926-01bcfa37eaad. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 938.820334] env[62814]: DEBUG nova.network.neutron [req-c009f7f9-dba0-49c2-a956-ec8bea241d24 req-578b5100-f612-4b98-a93d-ea0ab796ad2c service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updating instance_info_cache with network_info: [{"id": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "address": "fa:16:3e:aa:55:02", "network": {"id": "f638655e-3684-4eda-a001-01da430d0537", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-513788921-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "42a09e557c794726aa4895c1563ea62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9c621a9c-66f5-426a-8aab-bd8b2e912106", "external-id": "nsx-vlan-transportzone-485", "segmentation_id": 485, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap95b486fb-bb", "ovs_interfaceid": "95b486fb-bbe0-4666-a926-01bcfa37eaad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.887640] env[62814]: DEBUG nova.compute.manager [req-4b81f7e3-ad3d-487c-bb36-c4ed7456a960 req-1e7d88fa-ed68-4d07-952e-7deb7a5ffd8c service nova] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Received event network-changed-20070e44-989b-4d3c-aab6-3de905aad547 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 938.888138] env[62814]: DEBUG nova.compute.manager [req-4b81f7e3-ad3d-487c-bb36-c4ed7456a960 req-1e7d88fa-ed68-4d07-952e-7deb7a5ffd8c service nova] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Refreshing instance network info cache due to event network-changed-20070e44-989b-4d3c-aab6-3de905aad547. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 938.888138] env[62814]: DEBUG oslo_concurrency.lockutils [req-4b81f7e3-ad3d-487c-bb36-c4ed7456a960 req-1e7d88fa-ed68-4d07-952e-7deb7a5ffd8c service nova] Acquiring lock "refresh_cache-f54b8bb9-69b7-4bb4-a82c-9f796050e719" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.888307] env[62814]: DEBUG oslo_concurrency.lockutils [req-4b81f7e3-ad3d-487c-bb36-c4ed7456a960 req-1e7d88fa-ed68-4d07-952e-7deb7a5ffd8c service nova] Acquired lock "refresh_cache-f54b8bb9-69b7-4bb4-a82c-9f796050e719" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 938.888376] env[62814]: DEBUG nova.network.neutron [req-4b81f7e3-ad3d-487c-bb36-c4ed7456a960 req-1e7d88fa-ed68-4d07-952e-7deb7a5ffd8c service nova] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Refreshing network info cache for port 20070e44-989b-4d3c-aab6-3de905aad547 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.896572] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294037, 'name': CreateVM_Task, 'duration_secs': 0.514824} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.896935] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.897536] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.898494] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 938.898981] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 938.899437] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d5c6952-8d71-412c-91f2-f8ffcf7295c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.906474] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 938.906474] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52edf42f-ff0b-895c-ce3a-25cc1fd7f360" [ 938.906474] env[62814]: _type = "Task" [ 938.906474] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.916456] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52edf42f-ff0b-895c-ce3a-25cc1fd7f360, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.932385] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294034, 'name': ReconfigVM_Task, 'duration_secs': 0.642564} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.932600] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 98ba2786-023f-4d36-bdd0-e38bdd23d73c/98ba2786-023f-4d36-bdd0-e38bdd23d73c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.933471] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d41c0cb3-9972-4a0a-9540-9793c71bb584 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.940390] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 938.940390] env[62814]: value = "task-4294038" [ 938.940390] env[62814]: _type = "Task" [ 938.940390] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.955773] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294038, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.241173] env[62814]: DEBUG nova.scheduler.client.report [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 939.322780] env[62814]: DEBUG oslo_concurrency.lockutils [req-c009f7f9-dba0-49c2-a956-ec8bea241d24 req-578b5100-f612-4b98-a93d-ea0ab796ad2c service nova] Releasing lock "refresh_cache-3e4479a0-8edd-4b37-8cc9-2c91275b88ee" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 939.424247] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52edf42f-ff0b-895c-ce3a-25cc1fd7f360, 'name': SearchDatastore_Task, 'duration_secs': 0.027595} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.425431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 939.425689] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.425966] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.426137] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 939.426340] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.426643] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8e7facb-5667-49f8-9838-2ca37cd976aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.438182] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.438388] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 939.439358] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-546d9e40-7b53-4e8f-a236-c98af25a2fbb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.450587] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294038, 'name': Rename_Task, 'duration_secs': 0.198461} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.452255] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 939.452962] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 939.452962] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bb701e-b590-a68b-34fb-30ccd496876f" [ 939.452962] env[62814]: _type = "Task" [ 939.452962] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.453306] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edd0ba5e-69f1-40d2-abab-5a27103d5d05 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.463786] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bb701e-b590-a68b-34fb-30ccd496876f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.465130] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 939.465130] env[62814]: value = "task-4294039" [ 939.465130] env[62814]: _type = "Task" [ 939.465130] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.472982] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294039, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.561977] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.564020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.640268] env[62814]: DEBUG nova.network.neutron [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Successfully updated port: 5992ceda-be00-450a-a26e-10cab2e2a02b {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 939.694469] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.694753] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.694975] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 939.695187] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.695373] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 939.698566] env[62814]: INFO nova.compute.manager [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Terminating instance [ 939.748601] env[62814]: DEBUG oslo_concurrency.lockutils [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.255s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 939.751985] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.897s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 939.752383] env[62814]: DEBUG nova.objects.instance [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lazy-loading 'resources' on Instance uuid bd0933ca-aab7-4dd4-a570-1a58a720f377 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.785092] env[62814]: INFO nova.scheduler.client.report [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Deleted allocations for instance cd89e4d5-c08e-42c9-aea8-84c20714e196 [ 939.969215] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bb701e-b590-a68b-34fb-30ccd496876f, 'name': SearchDatastore_Task, 'duration_secs': 0.015332} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.972676] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dbd2840-2a94-46de-bc17-5b183549ff7f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.982432] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294039, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.984287] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 939.984287] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52736dad-04e6-08c7-1792-b535dee4a946" [ 939.984287] env[62814]: _type = "Task" [ 939.984287] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.985459] env[62814]: DEBUG nova.network.neutron [req-4b81f7e3-ad3d-487c-bb36-c4ed7456a960 req-1e7d88fa-ed68-4d07-952e-7deb7a5ffd8c service nova] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Updated VIF entry in instance network info cache for port 20070e44-989b-4d3c-aab6-3de905aad547. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.985674] env[62814]: DEBUG nova.network.neutron [req-4b81f7e3-ad3d-487c-bb36-c4ed7456a960 req-1e7d88fa-ed68-4d07-952e-7deb7a5ffd8c service nova] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Updating instance_info_cache with network_info: [{"id": "20070e44-989b-4d3c-aab6-3de905aad547", "address": "fa:16:3e:bb:93:a1", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.208", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20070e44-98", "ovs_interfaceid": "20070e44-989b-4d3c-aab6-3de905aad547", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.999363] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52736dad-04e6-08c7-1792-b535dee4a946, 'name': SearchDatastore_Task, 'duration_secs': 0.016016} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.000387] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 940.000694] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] f54b8bb9-69b7-4bb4-a82c-9f796050e719/f54b8bb9-69b7-4bb4-a82c-9f796050e719.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.002069] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d4e6de4-603a-4e73-9361-c8d4d6a2df8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.005615] env[62814]: DEBUG nova.compute.manager [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Received event network-vif-plugged-5992ceda-be00-450a-a26e-10cab2e2a02b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 940.005854] env[62814]: DEBUG oslo_concurrency.lockutils [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] Acquiring lock "508321ab-ce10-4953-a9e3-193b9975bec7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.006160] env[62814]: DEBUG oslo_concurrency.lockutils [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] Lock "508321ab-ce10-4953-a9e3-193b9975bec7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 940.006312] env[62814]: DEBUG oslo_concurrency.lockutils [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] Lock "508321ab-ce10-4953-a9e3-193b9975bec7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.006512] env[62814]: DEBUG nova.compute.manager [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] No waiting events found dispatching network-vif-plugged-5992ceda-be00-450a-a26e-10cab2e2a02b {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 940.006694] env[62814]: WARNING nova.compute.manager [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Received unexpected event network-vif-plugged-5992ceda-be00-450a-a26e-10cab2e2a02b for instance with vm_state building and task_state spawning. [ 940.006894] env[62814]: DEBUG nova.compute.manager [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Received event network-changed-5992ceda-be00-450a-a26e-10cab2e2a02b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 940.007098] env[62814]: DEBUG nova.compute.manager [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Refreshing instance network info cache due to event network-changed-5992ceda-be00-450a-a26e-10cab2e2a02b. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 940.007355] env[62814]: DEBUG oslo_concurrency.lockutils [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] Acquiring lock "refresh_cache-508321ab-ce10-4953-a9e3-193b9975bec7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.007481] env[62814]: DEBUG oslo_concurrency.lockutils [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] Acquired lock "refresh_cache-508321ab-ce10-4953-a9e3-193b9975bec7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 940.007668] env[62814]: DEBUG nova.network.neutron [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Refreshing network info cache for port 5992ceda-be00-450a-a26e-10cab2e2a02b {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.016197] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 940.016197] env[62814]: value = "task-4294040" [ 940.016197] env[62814]: _type = "Task" [ 940.016197] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.025659] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294040, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.068230] env[62814]: DEBUG nova.compute.manager [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 940.143259] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "refresh_cache-508321ab-ce10-4953-a9e3-193b9975bec7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.206039] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "refresh_cache-0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.206186] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquired lock "refresh_cache-0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 940.206384] env[62814]: DEBUG nova.network.neutron [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 940.295946] env[62814]: DEBUG oslo_concurrency.lockutils [None req-abd296fb-4508-4408-9f7b-6a7cd9c797e1 tempest-ServersV294TestFqdnHostnames-1245048135 tempest-ServersV294TestFqdnHostnames-1245048135-project-member] Lock "cd89e4d5-c08e-42c9-aea8-84c20714e196" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.439s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 940.477925] env[62814]: DEBUG oslo_vmware.api [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294039, 'name': PowerOnVM_Task, 'duration_secs': 0.755223} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.483175] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.483175] env[62814]: DEBUG nova.compute.manager [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 940.483175] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac635a5-b74e-400c-bfdc-6a25b745b66e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.494553] env[62814]: DEBUG oslo_concurrency.lockutils [req-4b81f7e3-ad3d-487c-bb36-c4ed7456a960 req-1e7d88fa-ed68-4d07-952e-7deb7a5ffd8c service nova] Releasing lock "refresh_cache-f54b8bb9-69b7-4bb4-a82c-9f796050e719" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 940.526601] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294040, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.577010] env[62814]: DEBUG nova.network.neutron [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 940.592675] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 940.657934] env[62814]: DEBUG nova.network.neutron [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.725720] env[62814]: DEBUG nova.network.neutron [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 940.790813] env[62814]: DEBUG nova.network.neutron [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.844019] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0129ab5f-8565-40fc-825f-6db3789a3b87 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.851300] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c752432-2571-45a8-bab1-a69836ad32aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.887473] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547f395d-990f-45d0-925e-04a19bd4901e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.895629] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b410ef-f9c6-40f8-bfde-224f9fc32fe1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.912250] env[62814]: DEBUG nova.compute.provider_tree [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.001622] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.033070] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294040, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.168071] env[62814]: DEBUG oslo_concurrency.lockutils [req-1faad469-ad22-43e5-a9a7-0e3475211055 req-55dff199-245b-409b-8ed8-6758a6a276b2 service nova] Releasing lock "refresh_cache-508321ab-ce10-4953-a9e3-193b9975bec7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 941.168071] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquired lock "refresh_cache-508321ab-ce10-4953-a9e3-193b9975bec7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 941.168071] env[62814]: DEBUG nova.network.neutron [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 941.294720] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Releasing lock "refresh_cache-0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 941.295142] env[62814]: DEBUG nova.compute.manager [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 941.295377] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 941.296422] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739b9c52-e676-4ce2-b6f9-944337e734c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.307206] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 941.307574] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95d7dbc1-705a-433c-bc5e-5fb60296494d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.315513] env[62814]: DEBUG oslo_vmware.api [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 941.315513] env[62814]: value = "task-4294041" [ 941.315513] env[62814]: _type = "Task" [ 941.315513] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.326066] env[62814]: DEBUG oslo_vmware.api [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294041, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.416374] env[62814]: DEBUG nova.scheduler.client.report [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 941.531206] env[62814]: DEBUG oslo_concurrency.lockutils [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "5c574786-64a4-4f07-a267-101ecaaa6938" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.531206] env[62814]: DEBUG oslo_concurrency.lockutils [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "5c574786-64a4-4f07-a267-101ecaaa6938" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 941.531454] env[62814]: DEBUG oslo_concurrency.lockutils [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "5c574786-64a4-4f07-a267-101ecaaa6938-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 941.531578] env[62814]: DEBUG oslo_concurrency.lockutils [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "5c574786-64a4-4f07-a267-101ecaaa6938-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 941.531741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "5c574786-64a4-4f07-a267-101ecaaa6938-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.533637] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294040, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.534484] env[62814]: INFO nova.compute.manager [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Terminating instance [ 941.700292] env[62814]: DEBUG nova.network.neutron [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 941.827215] env[62814]: DEBUG oslo_vmware.api [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294041, 'name': PowerOffVM_Task, 'duration_secs': 0.276565} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.827633] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.827742] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 941.828085] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1452d87-e7d4-4ba0-beb7-9b00037e1899 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.851493] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 941.851493] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 941.851973] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Deleting the datastore file [datastore2] 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.851973] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-71ad00be-e7cb-42dd-9fdb-4b316b575be6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.861223] env[62814]: DEBUG oslo_vmware.api [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for the task: (returnval){ [ 941.861223] env[62814]: value = "task-4294043" [ 941.861223] env[62814]: _type = "Task" [ 941.861223] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.868743] env[62814]: DEBUG nova.network.neutron [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Updating instance_info_cache with network_info: [{"id": "5992ceda-be00-450a-a26e-10cab2e2a02b", "address": "fa:16:3e:25:a6:3b", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5992ceda-be", "ovs_interfaceid": "5992ceda-be00-450a-a26e-10cab2e2a02b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.876057] env[62814]: DEBUG oslo_vmware.api [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294043, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.922758] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.171s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 941.925695] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.996s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 941.928625] env[62814]: INFO nova.compute.claims [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 941.959562] env[62814]: INFO nova.scheduler.client.report [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Deleted allocations for instance bd0933ca-aab7-4dd4-a570-1a58a720f377 [ 942.030878] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294040, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.588869} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.031201] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] f54b8bb9-69b7-4bb4-a82c-9f796050e719/f54b8bb9-69b7-4bb4-a82c-9f796050e719.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 942.031459] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.031767] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9947b645-eb0f-4231-865a-2c2e063eb164 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.038702] env[62814]: DEBUG nova.compute.manager [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 942.039031] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 942.039403] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 942.039403] env[62814]: value = "task-4294044" [ 942.039403] env[62814]: _type = "Task" [ 942.039403] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.040981] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1cbb1d-fc59-4e5f-9d96-1add0343f015 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.057754] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294044, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.060733] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.061185] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7832806e-e05f-4bec-b319-c8458ea47c6c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.068096] env[62814]: DEBUG oslo_vmware.api [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 942.068096] env[62814]: value = "task-4294045" [ 942.068096] env[62814]: _type = "Task" [ 942.068096] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.079120] env[62814]: DEBUG oslo_vmware.api [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294045, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.372249] env[62814]: DEBUG oslo_vmware.api [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Task: {'id': task-4294043, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17347} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.373014] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.373014] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 942.373014] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 942.373014] env[62814]: INFO nova.compute.manager [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Took 1.08 seconds to destroy the instance on the hypervisor. [ 942.373282] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 942.374950] env[62814]: DEBUG nova.compute.manager [-] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 942.374950] env[62814]: DEBUG nova.network.neutron [-] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 942.378909] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Releasing lock "refresh_cache-508321ab-ce10-4953-a9e3-193b9975bec7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 942.378909] env[62814]: DEBUG nova.compute.manager [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Instance network_info: |[{"id": "5992ceda-be00-450a-a26e-10cab2e2a02b", "address": "fa:16:3e:25:a6:3b", "network": {"id": "e119e86a-cec3-4f50-8dc2-faba83da80ef", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "32bba7a718ae4b3ba6a475c985d52bd7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a4f91f31-0516-4d62-a341-e03a50b7c477", "external-id": "nsx-vlan-transportzone-963", "segmentation_id": 963, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5992ceda-be", "ovs_interfaceid": "5992ceda-be00-450a-a26e-10cab2e2a02b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 942.379225] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:a6:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a4f91f31-0516-4d62-a341-e03a50b7c477', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5992ceda-be00-450a-a26e-10cab2e2a02b', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 942.387808] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 942.388112] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 942.388519] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-41eceeac-fea5-47e0-8fff-ab0e146d2b3e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.406137] env[62814]: DEBUG nova.network.neutron [-] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 942.413483] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 942.413483] env[62814]: value = "task-4294046" [ 942.413483] env[62814]: _type = "Task" [ 942.413483] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.423156] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294046, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.472888] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfac8e64-314e-483e-816e-f914b8bfc5d4 tempest-ServersTestBootFromVolume-1088202117 tempest-ServersTestBootFromVolume-1088202117-project-member] Lock "bd0933ca-aab7-4dd4-a570-1a58a720f377" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.105s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.558348] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294044, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110159} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.558820] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 942.560218] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47333cb7-b1eb-4682-9434-e26ed2ed8357 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.594213] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] f54b8bb9-69b7-4bb4-a82c-9f796050e719/f54b8bb9-69b7-4bb4-a82c-9f796050e719.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.597906] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f6848cf-d2d6-4431-9fe2-0a6541527a71 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.621785] env[62814]: DEBUG oslo_vmware.api [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294045, 'name': PowerOffVM_Task, 'duration_secs': 0.348524} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.623517] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 942.623625] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 942.624058] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 942.624058] env[62814]: value = "task-4294047" [ 942.624058] env[62814]: _type = "Task" [ 942.624058] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.624539] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50d5bade-5ddd-4903-b718-6111ea7e7a16 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.640067] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294047, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.698029] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 942.698029] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 942.698029] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Deleting the datastore file [datastore2] 5c574786-64a4-4f07-a267-101ecaaa6938 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.698029] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dad99e71-4cdd-4b3c-aaf8-0a42a5d1fe4c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.703390] env[62814]: DEBUG oslo_vmware.api [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for the task: (returnval){ [ 942.703390] env[62814]: value = "task-4294049" [ 942.703390] env[62814]: _type = "Task" [ 942.703390] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.711445] env[62814]: DEBUG oslo_vmware.api [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294049, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.909872] env[62814]: DEBUG nova.network.neutron [-] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.923454] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294046, 'name': CreateVM_Task, 'duration_secs': 0.441517} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.924308] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 942.925080] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.925269] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 942.925585] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 942.926159] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-880e08e6-88e9-4365-9f74-392428fc6293 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.931512] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 942.931512] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c167c8-392b-daa8-4c72-927e114e1f82" [ 942.931512] env[62814]: _type = "Task" [ 942.931512] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.932828] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "9be62576-5a05-473f-befd-b33f5fde9185" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.932828] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "9be62576-5a05-473f-befd-b33f5fde9185" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.932828] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "9be62576-5a05-473f-befd-b33f5fde9185-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 942.932828] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "9be62576-5a05-473f-befd-b33f5fde9185-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 942.932828] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "9be62576-5a05-473f-befd-b33f5fde9185-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 942.938662] env[62814]: INFO nova.compute.manager [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Terminating instance [ 942.951708] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c167c8-392b-daa8-4c72-927e114e1f82, 'name': SearchDatastore_Task, 'duration_secs': 0.016482} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.951931] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 942.952415] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 942.952488] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.952595] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 942.952789] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 942.953797] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dda78162-7ede-4a87-9c50-1897278e5b51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.965390] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 942.965838] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 942.970964] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9b30114-120f-48f3-a62a-46fd1cfb7627 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.977139] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 942.977139] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522dc747-45ee-cb7d-feb2-9044e1573d10" [ 942.977139] env[62814]: _type = "Task" [ 942.977139] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.986769] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522dc747-45ee-cb7d-feb2-9044e1573d10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.136568] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294047, 'name': ReconfigVM_Task, 'duration_secs': 0.446124} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.140970] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Reconfigured VM instance instance-00000046 to attach disk [datastore2] f54b8bb9-69b7-4bb4-a82c-9f796050e719/f54b8bb9-69b7-4bb4-a82c-9f796050e719.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.141940] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e48be41-28cb-4c50-8061-e8e3bf90cbf5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.150037] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 943.150037] env[62814]: value = "task-4294050" [ 943.150037] env[62814]: _type = "Task" [ 943.150037] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.165518] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294050, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.168846] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52081238-322c-735e-b8ae-0e13acb20987/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 943.170054] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37893568-62e5-4d18-9bdb-d951b6da048d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.176472] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52081238-322c-735e-b8ae-0e13acb20987/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 943.176661] env[62814]: ERROR oslo_vmware.rw_handles [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52081238-322c-735e-b8ae-0e13acb20987/disk-0.vmdk due to incomplete transfer. [ 943.179366] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-80044c81-4691-4571-8ff2-9eb959bdf45f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.187114] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52081238-322c-735e-b8ae-0e13acb20987/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 943.187734] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Uploaded image af88e617-86b6-4046-9983-8f685578152f to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 943.190243] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 943.190524] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-109611ca-9af7-4e4a-b2c2-ef5a74f4c95f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.198852] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 943.198852] env[62814]: value = "task-4294051" [ 943.198852] env[62814]: _type = "Task" [ 943.198852] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.208721] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294051, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.220646] env[62814]: DEBUG oslo_vmware.api [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Task: {'id': task-4294049, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.287513} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.220978] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.221189] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 943.221368] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 943.221540] env[62814]: INFO nova.compute.manager [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Took 1.18 seconds to destroy the instance on the hypervisor. [ 943.221833] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 943.222016] env[62814]: DEBUG nova.compute.manager [-] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 943.222115] env[62814]: DEBUG nova.network.neutron [-] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 943.413327] env[62814]: INFO nova.compute.manager [-] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Took 1.04 seconds to deallocate network for instance. [ 943.445670] env[62814]: DEBUG nova.compute.manager [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 943.446220] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 943.446920] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab4c801-8ecd-4edb-a66a-24be00c0bfa9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.457194] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 943.457655] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aadde4cf-60b2-4ef3-94c0-8937ad3aac41 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.465541] env[62814]: DEBUG oslo_vmware.api [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 943.465541] env[62814]: value = "task-4294052" [ 943.465541] env[62814]: _type = "Task" [ 943.465541] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.474111] env[62814]: DEBUG oslo_vmware.api [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4294052, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.478638] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8a1048-5f25-42a3-9795-85b2dd79ba80 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.491545] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56b6ede-c2e5-4454-a740-4040f5c1e917 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.494757] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522dc747-45ee-cb7d-feb2-9044e1573d10, 'name': SearchDatastore_Task, 'duration_secs': 0.01475} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.496330] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-556453c5-36c4-4f15-9bc5-35c80980a93c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.527205] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963a781f-c5c6-4ffe-8097-81ecbedfd6d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.531359] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 943.531359] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529bce43-d2d4-8220-8b5c-30f77a5ca55f" [ 943.531359] env[62814]: _type = "Task" [ 943.531359] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.538096] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f302e7-a7bc-479b-9cf0-e1d821c4b15b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.547554] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529bce43-d2d4-8220-8b5c-30f77a5ca55f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.559252] env[62814]: DEBUG nova.compute.provider_tree [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.611326] env[62814]: DEBUG nova.compute.manager [req-846e4a1d-238d-4963-80f1-16a78b3b5671 req-e104c219-b34d-4592-a697-35ad21bd6174 service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Received event network-vif-deleted-f3e3a6f1-ae61-48eb-a135-02afe3f22971 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 943.611326] env[62814]: INFO nova.compute.manager [req-846e4a1d-238d-4963-80f1-16a78b3b5671 req-e104c219-b34d-4592-a697-35ad21bd6174 service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Neutron deleted interface f3e3a6f1-ae61-48eb-a135-02afe3f22971; detaching it from the instance and deleting it from the info cache [ 943.611693] env[62814]: DEBUG nova.network.neutron [req-846e4a1d-238d-4963-80f1-16a78b3b5671 req-e104c219-b34d-4592-a697-35ad21bd6174 service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.660764] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294050, 'name': Rename_Task, 'duration_secs': 0.22842} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.660894] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 943.661369] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bfec6be7-7ffa-4974-965c-d84c9a182cb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.667786] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 943.667786] env[62814]: value = "task-4294053" [ 943.667786] env[62814]: _type = "Task" [ 943.667786] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.676169] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294053, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.709684] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294051, 'name': Destroy_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.921793] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 943.982551] env[62814]: DEBUG oslo_vmware.api [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4294052, 'name': PowerOffVM_Task, 'duration_secs': 0.355333} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.983444] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 943.983928] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 943.984355] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9bf516e8-1c5b-464f-84ef-b236714223b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.042687] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529bce43-d2d4-8220-8b5c-30f77a5ca55f, 'name': SearchDatastore_Task, 'duration_secs': 0.025871} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.043197] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 944.043631] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 508321ab-ce10-4953-a9e3-193b9975bec7/508321ab-ce10-4953-a9e3-193b9975bec7.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 944.046020] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cec786f3-9730-4688-a1fd-d5e36e7d22a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.048282] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 944.048702] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 944.049082] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Deleting the datastore file [datastore2] 9be62576-5a05-473f-befd-b33f5fde9185 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 944.049470] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad9cc009-af02-4c6a-ad11-c8c61fa86571 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.052520] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 944.052520] env[62814]: value = "task-4294055" [ 944.052520] env[62814]: _type = "Task" [ 944.052520] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.057936] env[62814]: DEBUG oslo_vmware.api [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for the task: (returnval){ [ 944.057936] env[62814]: value = "task-4294056" [ 944.057936] env[62814]: _type = "Task" [ 944.057936] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.065709] env[62814]: DEBUG nova.scheduler.client.report [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 944.069528] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294055, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.073762] env[62814]: DEBUG oslo_vmware.api [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4294056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.083090] env[62814]: DEBUG nova.network.neutron [-] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.114379] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-295cabf9-912e-47fc-9e73-af18aaf7233a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.123500] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ef4293-1db8-4b72-bc40-79a9be1abf2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.163477] env[62814]: DEBUG nova.compute.manager [req-846e4a1d-238d-4963-80f1-16a78b3b5671 req-e104c219-b34d-4592-a697-35ad21bd6174 service nova] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Detach interface failed, port_id=f3e3a6f1-ae61-48eb-a135-02afe3f22971, reason: Instance 5c574786-64a4-4f07-a267-101ecaaa6938 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 944.176915] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294053, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.209644] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294051, 'name': Destroy_Task, 'duration_secs': 0.66505} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.209644] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Destroyed the VM [ 944.209644] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 944.209804] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-46e67202-3b32-4104-b19b-a91d285675a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.216161] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 944.216161] env[62814]: value = "task-4294057" [ 944.216161] env[62814]: _type = "Task" [ 944.216161] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.223992] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294057, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.565386] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294055, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.572046] env[62814]: DEBUG oslo_vmware.api [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4294056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.572046] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.645s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 944.572046] env[62814]: DEBUG nova.compute.manager [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 944.574181] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.435s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 944.575596] env[62814]: INFO nova.compute.claims [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 944.584966] env[62814]: INFO nova.compute.manager [-] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Took 1.36 seconds to deallocate network for instance. [ 944.679200] env[62814]: DEBUG oslo_vmware.api [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294053, 'name': PowerOnVM_Task, 'duration_secs': 0.71948} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.679485] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 944.679689] env[62814]: INFO nova.compute.manager [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Took 9.01 seconds to spawn the instance on the hypervisor. [ 944.679870] env[62814]: DEBUG nova.compute.manager [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 944.680748] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32f7342-db6a-444f-9aba-bfbbe5770363 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.726032] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294057, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.072893] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294055, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.986071} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.078037] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 508321ab-ce10-4953-a9e3-193b9975bec7/508321ab-ce10-4953-a9e3-193b9975bec7.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 945.078450] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 945.078864] env[62814]: DEBUG oslo_vmware.api [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Task: {'id': task-4294056, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.948935} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.080477] env[62814]: DEBUG nova.compute.utils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 945.084789] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cfdc30df-f9b6-44b7-8ee7-2baeb058ba27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.089602] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 945.089917] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 945.090246] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 945.090526] env[62814]: INFO nova.compute.manager [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Took 1.64 seconds to destroy the instance on the hypervisor. [ 945.090903] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 945.092128] env[62814]: DEBUG nova.compute.manager [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 945.092412] env[62814]: DEBUG nova.network.neutron [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 945.095007] env[62814]: DEBUG nova.compute.manager [-] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 945.095161] env[62814]: DEBUG nova.network.neutron [-] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 945.098609] env[62814]: DEBUG oslo_concurrency.lockutils [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 945.109023] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 945.109023] env[62814]: value = "task-4294058" [ 945.109023] env[62814]: _type = "Task" [ 945.109023] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.120151] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294058, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.164185] env[62814]: DEBUG nova.policy [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd71116c0cfd4b509e4070ecc84613f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd82993ef7dfa4d2f8f39db3577dd321a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 945.200637] env[62814]: INFO nova.compute.manager [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Took 30.38 seconds to build instance. [ 945.227227] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294057, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.510601] env[62814]: DEBUG nova.network.neutron [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Successfully created port: 3c36df9c-47da-4470-9e9d-87659604c923 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 945.596906] env[62814]: DEBUG nova.compute.manager [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 945.617424] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294058, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.159271} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.617705] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 945.618504] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1163d2f9-0e8b-4b8a-a059-95694eb7e149 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.650483] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 508321ab-ce10-4953-a9e3-193b9975bec7/508321ab-ce10-4953-a9e3-193b9975bec7.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 945.655681] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7adce85-6a5d-4b12-a4a9-7fd5a0ec770d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.688024] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 945.688024] env[62814]: value = "task-4294059" [ 945.688024] env[62814]: _type = "Task" [ 945.688024] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.697970] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294059, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.706277] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3946782e-6ede-4d16-970e-960203046a55 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.899s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 945.732540] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294057, 'name': RemoveSnapshot_Task} progress is 36%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.123078] env[62814]: DEBUG nova.compute.manager [req-b2a0fd0f-4e0a-495f-95ca-a09a6379cf2e req-391028af-02b8-4ed0-99bc-1fde22b8b864 service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Received event network-vif-deleted-5ec47315-906b-4ad9-a8c8-87622cb99c27 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 946.123078] env[62814]: INFO nova.compute.manager [req-b2a0fd0f-4e0a-495f-95ca-a09a6379cf2e req-391028af-02b8-4ed0-99bc-1fde22b8b864 service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Neutron deleted interface 5ec47315-906b-4ad9-a8c8-87622cb99c27; detaching it from the instance and deleting it from the info cache [ 946.123078] env[62814]: DEBUG nova.network.neutron [req-b2a0fd0f-4e0a-495f-95ca-a09a6379cf2e req-391028af-02b8-4ed0-99bc-1fde22b8b864 service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.162528] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430fa9d3-5bd7-45aa-b8c8-62b0cdba7028 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.172195] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b46fc2-3648-403e-b035-3c3246a2e83e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.209767] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b4757a-f453-4006-9fbd-d26818b755e1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.217706] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.223712] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f431ebb-1cbd-44ea-9d13-c10fdbaad143 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.232081] env[62814]: DEBUG oslo_vmware.api [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294057, 'name': RemoveSnapshot_Task, 'duration_secs': 1.693809} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.239854] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 946.240121] env[62814]: INFO nova.compute.manager [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Took 18.41 seconds to snapshot the instance on the hypervisor. [ 946.243996] env[62814]: DEBUG nova.compute.provider_tree [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.598222] env[62814]: DEBUG nova.network.neutron [-] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.606888] env[62814]: DEBUG nova.compute.manager [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 946.626717] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bbdfdc19-bf1e-4269-80aa-cab821579815 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.632712] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 946.632958] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 946.633166] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 946.633364] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 946.633514] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 946.633662] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 946.633874] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 946.634047] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 946.634223] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 946.634378] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 946.634550] env[62814]: DEBUG nova.virt.hardware [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 946.635387] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31290b93-3988-4032-acb0-f1badfd4f40a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.642078] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e501ac3-986d-4d36-ab12-e2fef92b6f3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.656587] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02a1990-21a6-4a56-b526-d7471e694117 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.688324] env[62814]: DEBUG nova.compute.manager [req-b2a0fd0f-4e0a-495f-95ca-a09a6379cf2e req-391028af-02b8-4ed0-99bc-1fde22b8b864 service nova] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Detach interface failed, port_id=5ec47315-906b-4ad9-a8c8-87622cb99c27, reason: Instance 9be62576-5a05-473f-befd-b33f5fde9185 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 946.714700] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294059, 'name': ReconfigVM_Task, 'duration_secs': 0.900811} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.715007] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 508321ab-ce10-4953-a9e3-193b9975bec7/508321ab-ce10-4953-a9e3-193b9975bec7.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 946.715656] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed6c7611-c9fd-40d2-a2e2-ebf51caa3e22 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.722732] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 946.722732] env[62814]: value = "task-4294060" [ 946.722732] env[62814]: _type = "Task" [ 946.722732] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.730715] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294060, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.751410] env[62814]: DEBUG nova.scheduler.client.report [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 946.796520] env[62814]: DEBUG nova.compute.manager [None req-7081dc67-1353-4ad5-850c-fe487ce224f3 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Found 1 images (rotation: 2) {{(pid=62814) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5017}} [ 947.073996] env[62814]: DEBUG nova.compute.manager [req-8df6d8ab-def3-4904-a004-f30856c9dbb2 req-9db33281-a841-4f6c-9718-3b3acb227795 service nova] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Received event network-vif-plugged-3c36df9c-47da-4470-9e9d-87659604c923 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 947.073996] env[62814]: DEBUG oslo_concurrency.lockutils [req-8df6d8ab-def3-4904-a004-f30856c9dbb2 req-9db33281-a841-4f6c-9718-3b3acb227795 service nova] Acquiring lock "22429810-3ea7-4472-9a90-018d6efcba1d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 947.073996] env[62814]: DEBUG oslo_concurrency.lockutils [req-8df6d8ab-def3-4904-a004-f30856c9dbb2 req-9db33281-a841-4f6c-9718-3b3acb227795 service nova] Lock "22429810-3ea7-4472-9a90-018d6efcba1d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 947.074413] env[62814]: DEBUG oslo_concurrency.lockutils [req-8df6d8ab-def3-4904-a004-f30856c9dbb2 req-9db33281-a841-4f6c-9718-3b3acb227795 service nova] Lock "22429810-3ea7-4472-9a90-018d6efcba1d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 947.074413] env[62814]: DEBUG nova.compute.manager [req-8df6d8ab-def3-4904-a004-f30856c9dbb2 req-9db33281-a841-4f6c-9718-3b3acb227795 service nova] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] No waiting events found dispatching network-vif-plugged-3c36df9c-47da-4470-9e9d-87659604c923 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 947.074522] env[62814]: WARNING nova.compute.manager [req-8df6d8ab-def3-4904-a004-f30856c9dbb2 req-9db33281-a841-4f6c-9718-3b3acb227795 service nova] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Received unexpected event network-vif-plugged-3c36df9c-47da-4470-9e9d-87659604c923 for instance with vm_state building and task_state spawning. [ 947.100919] env[62814]: INFO nova.compute.manager [-] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Took 2.01 seconds to deallocate network for instance. [ 947.159132] env[62814]: DEBUG nova.network.neutron [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Successfully updated port: 3c36df9c-47da-4470-9e9d-87659604c923 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 947.233917] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294060, 'name': Rename_Task, 'duration_secs': 0.201258} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.234229] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 947.234495] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7694b4c-1ec0-4743-a27d-1506019d5c46 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.241926] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 947.241926] env[62814]: value = "task-4294061" [ 947.241926] env[62814]: _type = "Task" [ 947.241926] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.249650] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294061, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.256961] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.682s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 947.257050] env[62814]: DEBUG nova.compute.manager [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 947.260613] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.884s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 947.261461] env[62814]: DEBUG nova.objects.instance [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lazy-loading 'resources' on Instance uuid 98884571-8101-4ae9-b2e3-aeed97e3618f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.607998] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 947.662173] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "refresh_cache-22429810-3ea7-4472-9a90-018d6efcba1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.662385] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "refresh_cache-22429810-3ea7-4472-9a90-018d6efcba1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 947.662596] env[62814]: DEBUG nova.network.neutron [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 947.753101] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294061, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.764135] env[62814]: DEBUG nova.compute.utils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 947.768115] env[62814]: DEBUG nova.compute.manager [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 947.768290] env[62814]: DEBUG nova.network.neutron [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 947.843312] env[62814]: DEBUG nova.policy [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e246f49b0b84cd093549b6d6b384e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6e52480dd2c467790622901940cf385', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 948.229059] env[62814]: DEBUG nova.network.neutron [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 948.256024] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294061, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.267173] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c827d47-dfec-45dd-bb54-9cad07840b12 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.270352] env[62814]: DEBUG nova.compute.manager [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 948.278258] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe220a9e-8db8-4805-bdb2-80fcbd7fe7c1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.311571] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b4b7c7-0b1b-4ae6-b0f9-07351c389ed1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.320885] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57dcbf5-70b8-41af-b4c8-a164209530ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.326672] env[62814]: DEBUG nova.network.neutron [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Successfully created port: ef217261-6aa3-43f9-97b8-9c8d818e17f4 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.337773] env[62814]: DEBUG nova.compute.provider_tree [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.375040] env[62814]: DEBUG nova.compute.manager [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 948.378467] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca4450d-3480-4c03-8111-aa0feced463d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.428220] env[62814]: DEBUG nova.network.neutron [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Updating instance_info_cache with network_info: [{"id": "3c36df9c-47da-4470-9e9d-87659604c923", "address": "fa:16:3e:e7:74:b6", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c36df9c-47", "ovs_interfaceid": "3c36df9c-47da-4470-9e9d-87659604c923", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.756614] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294061, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.843450] env[62814]: DEBUG nova.scheduler.client.report [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 948.892032] env[62814]: INFO nova.compute.manager [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] instance snapshotting [ 948.892563] env[62814]: DEBUG nova.objects.instance [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'flavor' on Instance uuid 9343301c-dfe9-41b0-b4a0-067af544d297 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.930078] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "refresh_cache-22429810-3ea7-4472-9a90-018d6efcba1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 948.930410] env[62814]: DEBUG nova.compute.manager [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Instance network_info: |[{"id": "3c36df9c-47da-4470-9e9d-87659604c923", "address": "fa:16:3e:e7:74:b6", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c36df9c-47", "ovs_interfaceid": "3c36df9c-47da-4470-9e9d-87659604c923", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 948.930813] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:74:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c36df9c-47da-4470-9e9d-87659604c923', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 948.939069] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 948.939069] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 948.939069] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59bf8b17-3c66-42e4-842c-a4af22f5391b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.960051] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 948.960051] env[62814]: value = "task-4294062" [ 948.960051] env[62814]: _type = "Task" [ 948.960051] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.968622] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294062, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.197768] env[62814]: DEBUG nova.compute.manager [req-0e8c6a4a-f6b4-483b-9ae3-ebf546cd8b9d req-778dca35-7678-4e51-a504-725629b4a21e service nova] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Received event network-changed-3c36df9c-47da-4470-9e9d-87659604c923 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 949.198006] env[62814]: DEBUG nova.compute.manager [req-0e8c6a4a-f6b4-483b-9ae3-ebf546cd8b9d req-778dca35-7678-4e51-a504-725629b4a21e service nova] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Refreshing instance network info cache due to event network-changed-3c36df9c-47da-4470-9e9d-87659604c923. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 949.198233] env[62814]: DEBUG oslo_concurrency.lockutils [req-0e8c6a4a-f6b4-483b-9ae3-ebf546cd8b9d req-778dca35-7678-4e51-a504-725629b4a21e service nova] Acquiring lock "refresh_cache-22429810-3ea7-4472-9a90-018d6efcba1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.198377] env[62814]: DEBUG oslo_concurrency.lockutils [req-0e8c6a4a-f6b4-483b-9ae3-ebf546cd8b9d req-778dca35-7678-4e51-a504-725629b4a21e service nova] Acquired lock "refresh_cache-22429810-3ea7-4472-9a90-018d6efcba1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.198536] env[62814]: DEBUG nova.network.neutron [req-0e8c6a4a-f6b4-483b-9ae3-ebf546cd8b9d req-778dca35-7678-4e51-a504-725629b4a21e service nova] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Refreshing network info cache for port 3c36df9c-47da-4470-9e9d-87659604c923 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 949.255351] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294061, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.283435] env[62814]: DEBUG nova.compute.manager [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 949.309576] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 949.309849] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.310011] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 949.310214] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.310369] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 949.310512] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 949.310727] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 949.310926] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 949.311114] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 949.311286] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 949.311461] env[62814]: DEBUG nova.virt.hardware [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 949.312362] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a957e7d-aa70-421a-a54c-1db6f7aec0d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.320706] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a0dfa4-b6e4-4e07-ae16-db4742f1abe8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.349235] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.088s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.351622] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.869s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.351763] env[62814]: DEBUG nova.objects.instance [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lazy-loading 'resources' on Instance uuid d7332882-19b3-4ab3-8ea6-51d33c584844 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.375133] env[62814]: INFO nova.scheduler.client.report [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Deleted allocations for instance 98884571-8101-4ae9-b2e3-aeed97e3618f [ 949.398296] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85631112-cc16-4366-98e1-4273bf320262 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.416829] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279ac724-eaf8-432a-b757-2365b35991c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.472383] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294062, 'name': CreateVM_Task, 'duration_secs': 0.472665} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.472500] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 949.473217] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.473413] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.473735] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 949.474466] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d719550-0771-47f8-8361-39ee74116053 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.479464] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 949.479464] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52352a14-553d-73b7-7470-c4f94d55f70f" [ 949.479464] env[62814]: _type = "Task" [ 949.479464] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.487429] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52352a14-553d-73b7-7470-c4f94d55f70f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.757168] env[62814]: DEBUG oslo_vmware.api [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294061, 'name': PowerOnVM_Task, 'duration_secs': 2.179673} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.757360] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.757699] env[62814]: INFO nova.compute.manager [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Took 11.26 seconds to spawn the instance on the hypervisor. [ 949.757987] env[62814]: DEBUG nova.compute.manager [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 949.758840] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54656e97-c9a0-409e-938a-8cc19f7fd4ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.890385] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d0487dbe-d137-4b84-b5a8-4d871a55e0aa tempest-ServersAdmin275Test-185617392 tempest-ServersAdmin275Test-185617392-project-member] Lock "98884571-8101-4ae9-b2e3-aeed97e3618f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.860s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.927641] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 949.927966] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-75e422ba-18d0-46bf-a1d4-9255401000a2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.939488] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 949.939488] env[62814]: value = "task-4294063" [ 949.939488] env[62814]: _type = "Task" [ 949.939488] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.949453] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294063, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.966460] env[62814]: DEBUG nova.network.neutron [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Successfully updated port: ef217261-6aa3-43f9-97b8-9c8d818e17f4 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 949.970152] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquiring lock "844737d8-d852-44bb-bf9d-e673e737ef33" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.970415] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lock "844737d8-d852-44bb-bf9d-e673e737ef33" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.970626] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquiring lock "844737d8-d852-44bb-bf9d-e673e737ef33-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 949.971072] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lock "844737d8-d852-44bb-bf9d-e673e737ef33-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 949.971072] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lock "844737d8-d852-44bb-bf9d-e673e737ef33-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 949.974078] env[62814]: INFO nova.compute.manager [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Terminating instance [ 949.993503] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52352a14-553d-73b7-7470-c4f94d55f70f, 'name': SearchDatastore_Task, 'duration_secs': 0.010905} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.994187] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 949.994187] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 949.995175] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.995175] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 949.995175] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 949.995175] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-02fef6a9-f145-4f96-8aef-55766d6a3da7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.005333] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.005333] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.010041] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13511f56-29f5-4bfc-b5c2-d459ffac5d51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.016490] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 950.016490] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5242a5c0-1942-37df-ea3c-84e502503fc1" [ 950.016490] env[62814]: _type = "Task" [ 950.016490] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.027085] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5242a5c0-1942-37df-ea3c-84e502503fc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.036199] env[62814]: DEBUG nova.network.neutron [req-0e8c6a4a-f6b4-483b-9ae3-ebf546cd8b9d req-778dca35-7678-4e51-a504-725629b4a21e service nova] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Updated VIF entry in instance network info cache for port 3c36df9c-47da-4470-9e9d-87659604c923. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 950.036609] env[62814]: DEBUG nova.network.neutron [req-0e8c6a4a-f6b4-483b-9ae3-ebf546cd8b9d req-778dca35-7678-4e51-a504-725629b4a21e service nova] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Updating instance_info_cache with network_info: [{"id": "3c36df9c-47da-4470-9e9d-87659604c923", "address": "fa:16:3e:e7:74:b6", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c36df9c-47", "ovs_interfaceid": "3c36df9c-47da-4470-9e9d-87659604c923", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.274702] env[62814]: INFO nova.compute.manager [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Took 35.01 seconds to build instance. [ 950.367206] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e416e331-d9c0-47df-bcbb-aac5f558b9f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.377099] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ec025ff-db43-4c5f-957d-9be28f4246b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.411235] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b3041d-ce22-4a64-8de2-4cb3d2ad0a88 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.418599] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca2ada2-f21f-4a3f-8b0b-2b6300e9c3e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.431803] env[62814]: DEBUG nova.compute.provider_tree [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.449387] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294063, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.470651] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "refresh_cache-267c1ef2-ccc3-470e-943d-5fe9203c4dbc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.470830] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "refresh_cache-267c1ef2-ccc3-470e-943d-5fe9203c4dbc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 950.471112] env[62814]: DEBUG nova.network.neutron [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.479837] env[62814]: DEBUG nova.compute.manager [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 950.479837] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 950.480707] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfffa43e-d7ea-45d5-800d-2722b92bd809 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.489167] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 950.489414] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96b4aded-3cad-49db-8a54-ef6aea740b0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.495560] env[62814]: DEBUG oslo_vmware.api [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for the task: (returnval){ [ 950.495560] env[62814]: value = "task-4294064" [ 950.495560] env[62814]: _type = "Task" [ 950.495560] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.503703] env[62814]: DEBUG oslo_vmware.api [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4294064, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.527667] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5242a5c0-1942-37df-ea3c-84e502503fc1, 'name': SearchDatastore_Task, 'duration_secs': 0.009875} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.528482] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-170434f1-293d-498b-b148-a38deb075457 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.534640] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 950.534640] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529e9b3e-b282-805f-91f4-90b4c9deb2c6" [ 950.534640] env[62814]: _type = "Task" [ 950.534640] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.539655] env[62814]: DEBUG oslo_concurrency.lockutils [req-0e8c6a4a-f6b4-483b-9ae3-ebf546cd8b9d req-778dca35-7678-4e51-a504-725629b4a21e service nova] Releasing lock "refresh_cache-22429810-3ea7-4472-9a90-018d6efcba1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 950.544964] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529e9b3e-b282-805f-91f4-90b4c9deb2c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.776939] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8dfaa615-c124-44e3-99fc-91f9e36d386e tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "508321ab-ce10-4953-a9e3-193b9975bec7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.527s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 950.935803] env[62814]: DEBUG nova.scheduler.client.report [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 950.953909] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294063, 'name': CreateSnapshot_Task, 'duration_secs': 0.6184} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.954542] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 950.955563] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27df4e3-6904-4a80-b642-98c8ee026159 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.005222] env[62814]: DEBUG oslo_vmware.api [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4294064, 'name': PowerOffVM_Task, 'duration_secs': 0.247163} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.005536] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.005720] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 951.006025] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f6946eb-d3a6-4141-80cf-6e5144052ec5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.027210] env[62814]: DEBUG nova.network.neutron [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 951.045800] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529e9b3e-b282-805f-91f4-90b4c9deb2c6, 'name': SearchDatastore_Task, 'duration_secs': 0.017486} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.046460] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.046460] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 22429810-3ea7-4472-9a90-018d6efcba1d/22429810-3ea7-4472-9a90-018d6efcba1d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 951.047024] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-99bb143c-cb3f-4c6b-8d70-17b79d277682 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.057558] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 951.057558] env[62814]: value = "task-4294066" [ 951.057558] env[62814]: _type = "Task" [ 951.057558] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.067802] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294066, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.068696] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 951.068902] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 951.069091] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Deleting the datastore file [datastore2] 844737d8-d852-44bb-bf9d-e673e737ef33 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.069659] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6a25750f-3a2a-4618-aefc-8d36f5713078 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.076592] env[62814]: DEBUG oslo_vmware.api [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for the task: (returnval){ [ 951.076592] env[62814]: value = "task-4294067" [ 951.076592] env[62814]: _type = "Task" [ 951.076592] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.087323] env[62814]: DEBUG oslo_vmware.api [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4294067, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.222796] env[62814]: DEBUG nova.network.neutron [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Updating instance_info_cache with network_info: [{"id": "ef217261-6aa3-43f9-97b8-9c8d818e17f4", "address": "fa:16:3e:ea:fb:55", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef217261-6a", "ovs_interfaceid": "ef217261-6aa3-43f9-97b8-9c8d818e17f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.316515] env[62814]: DEBUG nova.compute.manager [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Received event network-vif-plugged-ef217261-6aa3-43f9-97b8-9c8d818e17f4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 951.316771] env[62814]: DEBUG oslo_concurrency.lockutils [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] Acquiring lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 951.317064] env[62814]: DEBUG oslo_concurrency.lockutils [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] Lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.317333] env[62814]: DEBUG oslo_concurrency.lockutils [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] Lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.317665] env[62814]: DEBUG nova.compute.manager [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] No waiting events found dispatching network-vif-plugged-ef217261-6aa3-43f9-97b8-9c8d818e17f4 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 951.317698] env[62814]: WARNING nova.compute.manager [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Received unexpected event network-vif-plugged-ef217261-6aa3-43f9-97b8-9c8d818e17f4 for instance with vm_state building and task_state spawning. [ 951.317867] env[62814]: DEBUG nova.compute.manager [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Received event network-changed-ef217261-6aa3-43f9-97b8-9c8d818e17f4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 951.318090] env[62814]: DEBUG nova.compute.manager [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Refreshing instance network info cache due to event network-changed-ef217261-6aa3-43f9-97b8-9c8d818e17f4. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 951.318251] env[62814]: DEBUG oslo_concurrency.lockutils [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] Acquiring lock "refresh_cache-267c1ef2-ccc3-470e-943d-5fe9203c4dbc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.443285] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.092s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 951.446759] env[62814]: DEBUG oslo_concurrency.lockutils [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.952s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 951.446759] env[62814]: DEBUG nova.objects.instance [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lazy-loading 'resources' on Instance uuid af6ef867-dc9a-4db6-8582-a6198e5caa77 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.479651] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 951.482244] env[62814]: INFO nova.scheduler.client.report [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Deleted allocations for instance d7332882-19b3-4ab3-8ea6-51d33c584844 [ 951.482975] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a44d09b0-62b0-4cae-aa16-0c417b873d1a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.498121] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 951.498121] env[62814]: value = "task-4294068" [ 951.498121] env[62814]: _type = "Task" [ 951.498121] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.510479] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294068, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.571655] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294066, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.589319] env[62814]: DEBUG oslo_vmware.api [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Task: {'id': task-4294067, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.233866} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.590463] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.590463] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 951.590463] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 951.590463] env[62814]: INFO nova.compute.manager [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Took 1.11 seconds to destroy the instance on the hypervisor. [ 951.590463] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 951.590691] env[62814]: DEBUG nova.compute.manager [-] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 951.590691] env[62814]: DEBUG nova.network.neutron [-] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 951.726560] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "refresh_cache-267c1ef2-ccc3-470e-943d-5fe9203c4dbc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 951.726560] env[62814]: DEBUG nova.compute.manager [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Instance network_info: |[{"id": "ef217261-6aa3-43f9-97b8-9c8d818e17f4", "address": "fa:16:3e:ea:fb:55", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef217261-6a", "ovs_interfaceid": "ef217261-6aa3-43f9-97b8-9c8d818e17f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 951.726560] env[62814]: DEBUG oslo_concurrency.lockutils [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] Acquired lock "refresh_cache-267c1ef2-ccc3-470e-943d-5fe9203c4dbc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 951.728817] env[62814]: DEBUG nova.network.neutron [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Refreshing network info cache for port ef217261-6aa3-43f9-97b8-9c8d818e17f4 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 951.729724] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ea:fb:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '51bac3c3-00ab-4a07-9e28-b3c951dee565', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef217261-6aa3-43f9-97b8-9c8d818e17f4', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.743402] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Creating folder: Project (e6e52480dd2c467790622901940cf385). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 951.746334] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d66e0b0-01ac-4916-b31a-50952cc68d9e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.758654] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Created folder: Project (e6e52480dd2c467790622901940cf385) in parent group-v845547. [ 951.758931] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Creating folder: Instances. Parent ref: group-v845753. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 951.759229] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c2789244-05da-4c09-bf76-40ebf4b60ac6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.772287] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Created folder: Instances in parent group-v845753. [ 951.772714] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 951.772968] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.773371] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0a4c3bd9-9056-4e81-8659-c63245dc4141 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.808028] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.808028] env[62814]: value = "task-4294071" [ 951.808028] env[62814]: _type = "Task" [ 951.808028] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.813911] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294071, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.008562] env[62814]: DEBUG oslo_concurrency.lockutils [None req-64c1bb60-3e0c-4af3-9fb8-f85d1d4ca11b tempest-ServersListShow298Test-825054976 tempest-ServersListShow298Test-825054976-project-member] Lock "d7332882-19b3-4ab3-8ea6-51d33c584844" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.872s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 952.030969] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294068, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.080501] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294066, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.689714} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.082198] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 22429810-3ea7-4472-9a90-018d6efcba1d/22429810-3ea7-4472-9a90-018d6efcba1d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.082459] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.086444] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fb70545b-1f84-4ee5-8d21-733004c9fe26 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.093870] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 952.093870] env[62814]: value = "task-4294072" [ 952.093870] env[62814]: _type = "Task" [ 952.093870] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.105079] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294072, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.327288] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294071, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.332378] env[62814]: DEBUG nova.compute.manager [req-d2289e7a-3672-4ca0-8e53-70e3dc78abf1 req-a5f19053-2b22-4ef4-a350-624bc73a9cdb service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Received event network-vif-deleted-5099a0d1-867c-4372-a319-ad384543c523 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 952.332590] env[62814]: INFO nova.compute.manager [req-d2289e7a-3672-4ca0-8e53-70e3dc78abf1 req-a5f19053-2b22-4ef4-a350-624bc73a9cdb service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Neutron deleted interface 5099a0d1-867c-4372-a319-ad384543c523; detaching it from the instance and deleting it from the info cache [ 952.332847] env[62814]: DEBUG nova.network.neutron [req-d2289e7a-3672-4ca0-8e53-70e3dc78abf1 req-a5f19053-2b22-4ef4-a350-624bc73a9cdb service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.502669] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5189462f-ccb3-4cd6-b130-a861a4232365 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.512465] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90074321-a740-4d51-98ed-97283e09c83a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.521708] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294068, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.550348] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc18d928-8bb4-4dcb-ac58-4e52d2bde4a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.557758] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6873144-ae1c-4f24-9c6d-b34674e38fcd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.573906] env[62814]: DEBUG nova.compute.provider_tree [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.603458] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294072, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075914} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.604727] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.604727] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2103df85-de08-47bc-98c8-a783caa1d572 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.628333] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 22429810-3ea7-4472-9a90-018d6efcba1d/22429810-3ea7-4472-9a90-018d6efcba1d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.628663] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5f2f3b3-27a3-40f3-8394-d392cee46707 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.649161] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 952.649161] env[62814]: value = "task-4294073" [ 952.649161] env[62814]: _type = "Task" [ 952.649161] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.660096] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294073, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.807431] env[62814]: DEBUG nova.network.neutron [-] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.818637] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294071, 'name': CreateVM_Task, 'duration_secs': 0.564608} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.818845] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 952.819713] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.819793] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 952.820072] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 952.821036] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75bf13d5-3f2a-4814-bfa1-aa7326d2fb37 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.829270] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 952.829270] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c20b0a-63ef-02fa-8854-8fd0e0bce08c" [ 952.829270] env[62814]: _type = "Task" [ 952.829270] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.841938] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c20b0a-63ef-02fa-8854-8fd0e0bce08c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.842378] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d98f934-ea9a-4981-9f4f-72779fe8ab37 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.853088] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45cd62d5-269e-49f7-9ac2-ddd0987a210c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.876497] env[62814]: DEBUG nova.compute.manager [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 952.897493] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9086ae93-798d-468a-9685-1101a3011232 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.901678] env[62814]: DEBUG nova.compute.manager [req-d2289e7a-3672-4ca0-8e53-70e3dc78abf1 req-a5f19053-2b22-4ef4-a350-624bc73a9cdb service nova] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Detach interface failed, port_id=5099a0d1-867c-4372-a319-ad384543c523, reason: Instance 844737d8-d852-44bb-bf9d-e673e737ef33 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 952.930157] env[62814]: DEBUG nova.network.neutron [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Updated VIF entry in instance network info cache for port ef217261-6aa3-43f9-97b8-9c8d818e17f4. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 952.930629] env[62814]: DEBUG nova.network.neutron [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Updating instance_info_cache with network_info: [{"id": "ef217261-6aa3-43f9-97b8-9c8d818e17f4", "address": "fa:16:3e:ea:fb:55", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef217261-6a", "ovs_interfaceid": "ef217261-6aa3-43f9-97b8-9c8d818e17f4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.021251] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294068, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.076994] env[62814]: DEBUG nova.scheduler.client.report [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 953.161789] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294073, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.313378] env[62814]: INFO nova.compute.manager [-] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Took 1.72 seconds to deallocate network for instance. [ 953.340094] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c20b0a-63ef-02fa-8854-8fd0e0bce08c, 'name': SearchDatastore_Task, 'duration_secs': 0.014964} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.343636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.343636] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.343636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.343636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 953.343636] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.343636] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc9679c7-be20-44a4-803c-527f1f6a458b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.352927] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.352927] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.352927] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8508d980-4b4e-455f-8d82-1e9eebb666c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.358628] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 953.358628] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528bc526-d84f-5c05-994a-87443bced67f" [ 953.358628] env[62814]: _type = "Task" [ 953.358628] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.366366] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528bc526-d84f-5c05-994a-87443bced67f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.410285] env[62814]: INFO nova.compute.manager [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] instance snapshotting [ 953.413189] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53f6cfc7-d1a8-4892-80a0-68352c9d2825 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.434587] env[62814]: DEBUG oslo_concurrency.lockutils [req-17cb7f4e-d8e0-41f5-bad9-43ac680b7ac0 req-3ea5dd15-33f7-4029-a34b-9cbedce5df02 service nova] Releasing lock "refresh_cache-267c1ef2-ccc3-470e-943d-5fe9203c4dbc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 953.439465] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15ae8b7-c34d-45a9-8434-0e6664072f90 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.516472] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294068, 'name': CloneVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.583114] env[62814]: DEBUG oslo_concurrency.lockutils [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.137s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 953.585739] env[62814]: DEBUG oslo_concurrency.lockutils [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.403s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 953.585968] env[62814]: DEBUG nova.objects.instance [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lazy-loading 'resources' on Instance uuid cbc5985a-38e7-4e52-9fb0-264b5cec013a {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.615852] env[62814]: INFO nova.scheduler.client.report [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted allocations for instance af6ef867-dc9a-4db6-8582-a6198e5caa77 [ 953.663066] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294073, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.820820] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 953.870505] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528bc526-d84f-5c05-994a-87443bced67f, 'name': SearchDatastore_Task, 'duration_secs': 0.027391} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.871889] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-201e0abe-176f-450c-87ae-482d2e5e8107 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.877849] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 953.877849] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f99b2c-9a11-95f4-dc9a-3bd7548688e9" [ 953.877849] env[62814]: _type = "Task" [ 953.877849] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.885251] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f99b2c-9a11-95f4-dc9a-3bd7548688e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.946757] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 953.947126] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-800be2d8-b3d6-43b1-af52-1cd6a74ef818 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.955472] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 953.955472] env[62814]: value = "task-4294074" [ 953.955472] env[62814]: _type = "Task" [ 953.955472] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.966903] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294074, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.018292] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294068, 'name': CloneVM_Task, 'duration_secs': 2.072747} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.018292] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Created linked-clone VM from snapshot [ 954.019122] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a647807c-e390-49c1-8f6f-eeda260081b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.026496] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Uploading image a300ce7b-810e-417d-a7cf-6a0c7af5239e {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 954.071307] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 954.071307] env[62814]: value = "vm-845752" [ 954.071307] env[62814]: _type = "VirtualMachine" [ 954.071307] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 954.071307] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a9a20979-de6e-4d36-b594-5cf3c7dd92e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.071307] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lease: (returnval){ [ 954.071307] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52408e11-af8c-3f6f-622f-042e9f33f6be" [ 954.071307] env[62814]: _type = "HttpNfcLease" [ 954.071307] env[62814]: } obtained for exporting VM: (result){ [ 954.071307] env[62814]: value = "vm-845752" [ 954.071307] env[62814]: _type = "VirtualMachine" [ 954.071307] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 954.071307] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the lease: (returnval){ [ 954.071307] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52408e11-af8c-3f6f-622f-042e9f33f6be" [ 954.071307] env[62814]: _type = "HttpNfcLease" [ 954.071307] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 954.071307] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.071307] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52408e11-af8c-3f6f-622f-042e9f33f6be" [ 954.071307] env[62814]: _type = "HttpNfcLease" [ 954.071307] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 954.126422] env[62814]: DEBUG oslo_concurrency.lockutils [None req-52b74d58-2f26-426b-9761-5f3ade36d608 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "af6ef867-dc9a-4db6-8582-a6198e5caa77" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.090s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.164220] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294073, 'name': ReconfigVM_Task, 'duration_secs': 1.406517} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.166842] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 22429810-3ea7-4472-9a90-018d6efcba1d/22429810-3ea7-4472-9a90-018d6efcba1d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 954.167736] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f8bdd50-5e7d-423e-b39e-37f4a0655506 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.176147] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 954.176147] env[62814]: value = "task-4294076" [ 954.176147] env[62814]: _type = "Task" [ 954.176147] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.188869] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294076, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.390334] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f99b2c-9a11-95f4-dc9a-3bd7548688e9, 'name': SearchDatastore_Task, 'duration_secs': 0.011199} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.390334] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 954.394368] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 267c1ef2-ccc3-470e-943d-5fe9203c4dbc/267c1ef2-ccc3-470e-943d-5fe9203c4dbc.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.394368] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64fa0668-fac2-4f0a-8a50-8d22676953c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.401144] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 954.401144] env[62814]: value = "task-4294077" [ 954.401144] env[62814]: _type = "Task" [ 954.401144] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.410419] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294077, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.467917] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294074, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.570297] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce68c3bf-6460-4e65-8907-e09cf8193252 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.576679] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.576679] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52408e11-af8c-3f6f-622f-042e9f33f6be" [ 954.576679] env[62814]: _type = "HttpNfcLease" [ 954.576679] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 954.579140] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 954.579140] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52408e11-af8c-3f6f-622f-042e9f33f6be" [ 954.579140] env[62814]: _type = "HttpNfcLease" [ 954.579140] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 954.580286] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b4d9fe2-191e-4ed0-a2f0-89493ce75a1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.583786] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5717b1d-f6dc-4d0f-a52c-ddcb2bcfc7b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.595021] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bec50-346a-83d1-620a-b17e3e4ce3e4/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 954.595262] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bec50-346a-83d1-620a-b17e3e4ce3e4/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 954.630096] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "c8765756-0870-4a06-a1a5-d02177959b29" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.630307] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "c8765756-0870-4a06-a1a5-d02177959b29" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.630521] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "c8765756-0870-4a06-a1a5-d02177959b29-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 954.630736] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "c8765756-0870-4a06-a1a5-d02177959b29-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 954.630982] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "c8765756-0870-4a06-a1a5-d02177959b29-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 954.636887] env[62814]: INFO nova.compute.manager [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Terminating instance [ 954.639038] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3128e411-fc9d-49ce-8a5b-773e74fa4aed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.712187] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9030a443-d985-473a-b73a-c460b57da2d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.721634] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294076, 'name': Rename_Task, 'duration_secs': 0.208011} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.722849] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 954.722849] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ebc4353e-dbba-41ee-a5de-acbf71378e14 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.736028] env[62814]: DEBUG nova.compute.provider_tree [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.744029] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 954.744029] env[62814]: value = "task-4294078" [ 954.744029] env[62814]: _type = "Task" [ 954.744029] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.752721] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294078, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.914866] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6c7052ab-6953-461d-942d-cf29a9894d92 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.919846] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294077, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.977693] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294074, 'name': CreateSnapshot_Task, 'duration_secs': 0.86612} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.978903] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 954.979786] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58d5375c-2480-4c15-98bd-757caac9c076 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.211947] env[62814]: DEBUG nova.compute.manager [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 955.211947] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.214447] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0e45189-233f-4d54-9584-026a5c69532f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.220197] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "fd1867e3-2523-4969-a157-b14c650f3779" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.220197] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "fd1867e3-2523-4969-a157-b14c650f3779" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 955.224356] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.224626] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01b43076-4ea5-4d49-a500-dad0d8309c0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.230997] env[62814]: DEBUG oslo_vmware.api [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 955.230997] env[62814]: value = "task-4294079" [ 955.230997] env[62814]: _type = "Task" [ 955.230997] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.238629] env[62814]: DEBUG nova.scheduler.client.report [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 955.244838] env[62814]: DEBUG oslo_vmware.api [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294079, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.258276] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294078, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.412858] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294077, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531131} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.413144] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 267c1ef2-ccc3-470e-943d-5fe9203c4dbc/267c1ef2-ccc3-470e-943d-5fe9203c4dbc.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 955.413477] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.414896] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-061e0c83-2bd4-4fc1-9ed4-eb38defa615a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.425017] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 955.425017] env[62814]: value = "task-4294080" [ 955.425017] env[62814]: _type = "Task" [ 955.425017] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.441438] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294080, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.501830] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 955.502285] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3685c456-0153-4b40-b047-a86703e63433 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.512464] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 955.512464] env[62814]: value = "task-4294081" [ 955.512464] env[62814]: _type = "Task" [ 955.512464] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.526305] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294081, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.722720] env[62814]: DEBUG nova.compute.manager [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 955.745183] env[62814]: DEBUG oslo_vmware.api [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294079, 'name': PowerOffVM_Task, 'duration_secs': 0.203621} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.746250] env[62814]: DEBUG oslo_concurrency.lockutils [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.160s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 955.748868] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 955.752428] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 955.759802] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.578s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 955.760132] env[62814]: DEBUG nova.objects.instance [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lazy-loading 'resources' on Instance uuid 99864c68-5f11-4119-b8e3-3aa8719f267b {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 955.761852] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dd51fe30-f346-4ee0-be5b-d465a11033f7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.772021] env[62814]: DEBUG oslo_vmware.api [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294078, 'name': PowerOnVM_Task, 'duration_secs': 0.729534} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.772407] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 955.773201] env[62814]: INFO nova.compute.manager [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Took 9.17 seconds to spawn the instance on the hypervisor. [ 955.773201] env[62814]: DEBUG nova.compute.manager [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 955.773976] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb50f7d-3c88-4a63-81f3-dc287a0a35d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.796714] env[62814]: INFO nova.scheduler.client.report [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleted allocations for instance cbc5985a-38e7-4e52-9fb0-264b5cec013a [ 955.828909] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 955.829475] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 955.829856] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleting the datastore file [datastore2] c8765756-0870-4a06-a1a5-d02177959b29 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.832252] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65b18d8c-678d-4770-96b6-34032be4239b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.844157] env[62814]: DEBUG oslo_vmware.api [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 955.844157] env[62814]: value = "task-4294083" [ 955.844157] env[62814]: _type = "Task" [ 955.844157] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.856350] env[62814]: DEBUG oslo_vmware.api [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294083, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.938877] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294080, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092165} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.939448] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.940674] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-470eef34-b9e8-45e9-8e34-e140113215a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.968233] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 267c1ef2-ccc3-470e-943d-5fe9203c4dbc/267c1ef2-ccc3-470e-943d-5fe9203c4dbc.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.970296] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8faeebfc-c7e2-425c-9c49-fa9ea0d9997b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.994952] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 955.995741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 956.007190] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 956.007190] env[62814]: value = "task-4294084" [ 956.007190] env[62814]: _type = "Task" [ 956.007190] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.026636] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294084, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.031660] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294081, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.254353] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 956.306636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-41edf308-fdbe-420b-b144-1b2d9c639a0d tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "cbc5985a-38e7-4e52-9fb0-264b5cec013a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.186s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 956.308484] env[62814]: INFO nova.compute.manager [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Took 38.40 seconds to build instance. [ 956.360533] env[62814]: DEBUG oslo_vmware.api [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294083, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29149} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.361011] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.361365] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.361701] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.362581] env[62814]: INFO nova.compute.manager [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Took 1.15 seconds to destroy the instance on the hypervisor. [ 956.362581] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 956.364060] env[62814]: DEBUG nova.compute.manager [-] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 956.364295] env[62814]: DEBUG nova.network.neutron [-] [instance: c8765756-0870-4a06-a1a5-d02177959b29] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.501221] env[62814]: DEBUG nova.compute.manager [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 956.525374] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294084, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.531113] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294081, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.744706] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d52a472-b882-486a-bce7-c2a498e93900 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.753373] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4161cc3b-e668-4c75-b937-d0b883368aec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.792033] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c12b982f-ad21-4d36-aa64-ead00a8d4500 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.798543] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bca3cea6-e913-4a4f-a91e-b99a3f9dae47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.805908] env[62814]: DEBUG nova.compute.manager [req-cd4bb304-cdf9-4996-b341-cac90da8c4fd req-ac9f8a06-28c6-440c-a8ea-5d53d11ffca1 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Received event network-vif-deleted-fa04dc3a-1732-4479-80e7-1d3a547ea8c5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 956.806059] env[62814]: INFO nova.compute.manager [req-cd4bb304-cdf9-4996-b341-cac90da8c4fd req-ac9f8a06-28c6-440c-a8ea-5d53d11ffca1 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Neutron deleted interface fa04dc3a-1732-4479-80e7-1d3a547ea8c5; detaching it from the instance and deleting it from the info cache [ 956.806307] env[62814]: DEBUG nova.network.neutron [req-cd4bb304-cdf9-4996-b341-cac90da8c4fd req-ac9f8a06-28c6-440c-a8ea-5d53d11ffca1 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.818654] env[62814]: DEBUG nova.compute.provider_tree [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.821087] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c4ba08d-7d26-42f7-a5f5-d36256657289 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "22429810-3ea7-4472-9a90-018d6efcba1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.469s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.019996] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294084, 'name': ReconfigVM_Task, 'duration_secs': 0.807887} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.023625] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 267c1ef2-ccc3-470e-943d-5fe9203c4dbc/267c1ef2-ccc3-470e-943d-5fe9203c4dbc.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.024593] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42e4b3ff-af85-43db-9282-493ad1f9380d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.031255] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294081, 'name': CloneVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.033519] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.033900] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 957.033900] env[62814]: value = "task-4294085" [ 957.033900] env[62814]: _type = "Task" [ 957.033900] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.042317] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294085, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.194239] env[62814]: DEBUG nova.network.neutron [-] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.248778] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.248778] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.248940] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.249128] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.249340] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.253016] env[62814]: INFO nova.compute.manager [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Terminating instance [ 957.325781] env[62814]: DEBUG nova.scheduler.client.report [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 957.329667] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf37b7a9-8e7d-4208-a9c6-f55c0c356a8f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.342794] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81129051-5306-4fa4-a5f4-4c0cb27e3870 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.388548] env[62814]: DEBUG nova.compute.manager [req-cd4bb304-cdf9-4996-b341-cac90da8c4fd req-ac9f8a06-28c6-440c-a8ea-5d53d11ffca1 service nova] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Detach interface failed, port_id=fa04dc3a-1732-4479-80e7-1d3a547ea8c5, reason: Instance c8765756-0870-4a06-a1a5-d02177959b29 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 957.530547] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294081, 'name': CloneVM_Task, 'duration_secs': 1.566527} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.530894] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Created linked-clone VM from snapshot [ 957.531519] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e597ec2-dedf-4d28-834b-272e5bdc83a3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.541429] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Uploading image f876a8b9-10c0-49c4-8d83-20d0d8b591c4 {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 957.547887] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294085, 'name': Rename_Task, 'duration_secs': 0.171965} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.548146] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 957.548374] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1037f3e-0b3a-4923-b3bd-e68750c49030 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.555596] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 957.555596] env[62814]: value = "task-4294086" [ 957.555596] env[62814]: _type = "Task" [ 957.555596] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.564020] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 957.564020] env[62814]: value = "vm-845757" [ 957.564020] env[62814]: _type = "VirtualMachine" [ 957.564020] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 957.564291] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-48ad3727-7fb1-48c2-a0ff-33f98ed0ab4f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.569242] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294086, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.575013] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lease: (returnval){ [ 957.575013] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dab574-3f35-bb9e-e9f3-36839955bd65" [ 957.575013] env[62814]: _type = "HttpNfcLease" [ 957.575013] env[62814]: } obtained for exporting VM: (result){ [ 957.575013] env[62814]: value = "vm-845757" [ 957.575013] env[62814]: _type = "VirtualMachine" [ 957.575013] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 957.575309] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the lease: (returnval){ [ 957.575309] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dab574-3f35-bb9e-e9f3-36839955bd65" [ 957.575309] env[62814]: _type = "HttpNfcLease" [ 957.575309] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 957.582218] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 957.582218] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dab574-3f35-bb9e-e9f3-36839955bd65" [ 957.582218] env[62814]: _type = "HttpNfcLease" [ 957.582218] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 957.698226] env[62814]: INFO nova.compute.manager [-] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Took 1.33 seconds to deallocate network for instance. [ 957.760023] env[62814]: DEBUG nova.compute.manager [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 957.760023] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 957.760323] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca6e1c9-f298-4975-ac39-a6076dd40b61 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.769233] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.769505] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ccaab5a-feb3-4803-a484-bb1dbc4b8ed0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.778248] env[62814]: DEBUG oslo_vmware.api [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 957.778248] env[62814]: value = "task-4294088" [ 957.778248] env[62814]: _type = "Task" [ 957.778248] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.791514] env[62814]: DEBUG oslo_vmware.api [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4294088, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.831073] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "22429810-3ea7-4472-9a90-018d6efcba1d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.831419] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "22429810-3ea7-4472-9a90-018d6efcba1d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.831744] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "22429810-3ea7-4472-9a90-018d6efcba1d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 957.831933] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "22429810-3ea7-4472-9a90-018d6efcba1d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.832160] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "22429810-3ea7-4472-9a90-018d6efcba1d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.834396] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.075s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 957.837112] env[62814]: INFO nova.compute.manager [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Terminating instance [ 957.839174] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.913s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 957.839174] env[62814]: DEBUG nova.objects.instance [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lazy-loading 'resources' on Instance uuid 425e8edd-c002-45a1-bb6f-ee3ac8812509 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.863698] env[62814]: INFO nova.scheduler.client.report [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Deleted allocations for instance 99864c68-5f11-4119-b8e3-3aa8719f267b [ 958.069679] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294086, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.084455] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 958.084455] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dab574-3f35-bb9e-e9f3-36839955bd65" [ 958.084455] env[62814]: _type = "HttpNfcLease" [ 958.084455] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 958.085678] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 958.085678] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dab574-3f35-bb9e-e9f3-36839955bd65" [ 958.085678] env[62814]: _type = "HttpNfcLease" [ 958.085678] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 958.085678] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f498a4-b09d-42bf-932c-be330cbb4bc3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.094629] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52396f16-108b-5b64-8ecd-5e633af20030/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 958.094810] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52396f16-108b-5b64-8ecd-5e633af20030/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 958.198407] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3c8eb0f2-3b77-4c40-8bc0-debaa762f1f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.204694] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 958.300158] env[62814]: DEBUG oslo_vmware.api [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4294088, 'name': PowerOffVM_Task, 'duration_secs': 0.293062} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.300158] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 958.300158] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 958.300158] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-598e89bb-127c-49f8-8c38-f844a284353b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.345176] env[62814]: DEBUG nova.compute.manager [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 958.345542] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 958.346470] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f79e64-5acf-43e5-a5c4-6213deb61f3c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.358042] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 958.358042] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbd9fa26-ef40-4f3c-a8f1-58a546b17506 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.366271] env[62814]: DEBUG oslo_vmware.api [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 958.366271] env[62814]: value = "task-4294090" [ 958.366271] env[62814]: _type = "Task" [ 958.366271] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.376650] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 958.376825] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 958.377022] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleting the datastore file [datastore2] f5ad9d70-75fb-4881-8853-5ede4d0903f2 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.378123] env[62814]: DEBUG oslo_concurrency.lockutils [None req-34d72cb0-4fb3-41b1-bdbc-e7282992ba83 tempest-ImagesOneServerNegativeTestJSON-1963370294 tempest-ImagesOneServerNegativeTestJSON-1963370294-project-member] Lock "99864c68-5f11-4119-b8e3-3aa8719f267b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.622s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 958.379120] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c21456e3-1cd5-4dda-b8a0-9e9379ceb33e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.385447] env[62814]: DEBUG oslo_vmware.api [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294090, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.394677] env[62814]: DEBUG oslo_vmware.api [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 958.394677] env[62814]: value = "task-4294091" [ 958.394677] env[62814]: _type = "Task" [ 958.394677] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.406061] env[62814]: DEBUG oslo_vmware.api [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4294091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.573441] env[62814]: DEBUG oslo_vmware.api [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294086, 'name': PowerOnVM_Task, 'duration_secs': 0.645427} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.573946] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 958.574208] env[62814]: INFO nova.compute.manager [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Took 9.29 seconds to spawn the instance on the hypervisor. [ 958.574466] env[62814]: DEBUG nova.compute.manager [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 958.575613] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc54682-9dd8-4791-b92e-f4ab07ad126c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.883898] env[62814]: DEBUG oslo_vmware.api [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294090, 'name': PowerOffVM_Task, 'duration_secs': 0.273855} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.887036] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 958.887036] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 958.887036] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-36825e1c-543b-436c-8686-9bac93fdfa04 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.889420] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538cf085-1fb1-474c-a6b1-4f9bdab58f46 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.901493] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d371e4-c571-43b5-bc03-97cdd68048c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.919082] env[62814]: DEBUG oslo_vmware.api [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4294091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.282945} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.951406] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 958.951830] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 958.952167] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 958.952505] env[62814]: INFO nova.compute.manager [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Took 1.19 seconds to destroy the instance on the hypervisor. [ 958.953023] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 958.953626] env[62814]: DEBUG nova.compute.manager [-] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 958.953956] env[62814]: DEBUG nova.network.neutron [-] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 958.956645] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08477252-509e-46d8-a1ff-007bc1b647b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.966106] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 958.966517] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 958.966921] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleting the datastore file [datastore2] 22429810-3ea7-4472-9a90-018d6efcba1d {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.969856] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84588721-577c-4512-9be2-a489d64271a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.973859] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-210c66a6-db64-4da1-9ebf-17cff7539b29 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.996140] env[62814]: DEBUG nova.compute.provider_tree [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.002023] env[62814]: DEBUG oslo_vmware.api [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 959.002023] env[62814]: value = "task-4294093" [ 959.002023] env[62814]: _type = "Task" [ 959.002023] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.102587] env[62814]: INFO nova.compute.manager [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Took 39.98 seconds to build instance. [ 959.501073] env[62814]: DEBUG nova.scheduler.client.report [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 959.517332] env[62814]: DEBUG oslo_vmware.api [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294093, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259984} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.518666] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.518666] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 959.518666] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 959.518666] env[62814]: INFO nova.compute.manager [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Took 1.17 seconds to destroy the instance on the hypervisor. [ 959.519889] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 959.520496] env[62814]: DEBUG nova.compute.manager [-] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 959.520728] env[62814]: DEBUG nova.network.neutron [-] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 959.605131] env[62814]: DEBUG oslo_concurrency.lockutils [None req-870fea88-aab6-4fb1-b4d0-f4f45f62d526 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.494s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 959.792435] env[62814]: DEBUG nova.network.neutron [-] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.012770] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.174s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.015719] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.363s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 960.016026] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.018319] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.966s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 960.018512] env[62814]: DEBUG nova.objects.instance [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 960.041359] env[62814]: INFO nova.scheduler.client.report [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Deleted allocations for instance 425e8edd-c002-45a1-bb6f-ee3ac8812509 [ 960.043692] env[62814]: INFO nova.scheduler.client.report [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Deleted allocations for instance 9cd89da4-d190-4bfc-81e1-da15b98d54ba [ 960.217479] env[62814]: DEBUG nova.network.neutron [-] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.295784] env[62814]: INFO nova.compute.manager [-] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Took 1.34 seconds to deallocate network for instance. [ 960.553614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8c07d47c-4729-491c-8279-2fe9255ae28a tempest-SecurityGroupsTestJSON-25847008 tempest-SecurityGroupsTestJSON-25847008-project-member] Lock "425e8edd-c002-45a1-bb6f-ee3ac8812509" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.017s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.555058] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4a44da1-128f-42a0-8090-ef6277b14478 tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "9cd89da4-d190-4bfc-81e1-da15b98d54ba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.678s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 960.720675] env[62814]: INFO nova.compute.manager [-] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Took 1.20 seconds to deallocate network for instance. [ 960.804245] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 961.029462] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0807a5d3-bf9d-4a44-bc29-5cdbb2aaa8f2 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 961.030681] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.446s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 961.032691] env[62814]: INFO nova.compute.claims [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.227444] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 962.370490] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-724439df-866b-4d2e-9c8f-a54dadeb8aaf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.380881] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108a1870-e7d9-400f-9b3b-7f69cde6eb75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.413555] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44407735-72f7-4d9c-adb6-223ac417b358 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.422762] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a92c349-4254-4b27-ba90-058bb6d13766 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.437785] env[62814]: DEBUG nova.compute.provider_tree [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.729975] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bec50-346a-83d1-620a-b17e3e4ce3e4/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 962.731104] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08686187-7e31-45c6-8e0d-347f6be34cc5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.739414] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bec50-346a-83d1-620a-b17e3e4ce3e4/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 962.739654] env[62814]: ERROR oslo_vmware.rw_handles [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bec50-346a-83d1-620a-b17e3e4ce3e4/disk-0.vmdk due to incomplete transfer. [ 962.739953] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c2a69b5a-3f8b-41e9-9b61-6d23f77cdc5d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.751572] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/526bec50-346a-83d1-620a-b17e3e4ce3e4/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 962.751572] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Uploaded image a300ce7b-810e-417d-a7cf-6a0c7af5239e to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 962.753813] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 962.754120] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-a0e3de05-2367-4921-b78d-0a7a11311d47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.763696] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 962.763696] env[62814]: value = "task-4294094" [ 962.763696] env[62814]: _type = "Task" [ 962.763696] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.773951] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294094, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.866118] env[62814]: DEBUG nova.compute.manager [req-878cdb1a-7644-4cfa-a19c-2c20e940f102 req-b3328339-bb30-415f-a3aa-0004ecdf98d5 service nova] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Received event network-vif-deleted-44e24b3d-908f-4ded-8f46-262fb433c4a6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 962.941281] env[62814]: DEBUG nova.scheduler.client.report [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 963.275440] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294094, 'name': Destroy_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.280081] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 963.280392] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 963.280604] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 963.280792] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 963.280980] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 963.283244] env[62814]: INFO nova.compute.manager [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Terminating instance [ 963.448628] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 963.449226] env[62814]: DEBUG nova.compute.manager [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 963.457438] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.865s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 963.459675] env[62814]: INFO nova.compute.claims [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 963.552299] env[62814]: DEBUG nova.objects.instance [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lazy-loading 'flavor' on Instance uuid d366a755-49b4-427b-8564-d8572a7fbbb7 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 963.786499] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294094, 'name': Destroy_Task, 'duration_secs': 0.625807} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.786499] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Destroyed the VM [ 963.786723] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 963.788024] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a0e987ac-ef6c-4957-9ff9-6287be9d314e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.789493] env[62814]: DEBUG nova.compute.manager [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 963.789779] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 963.790715] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221cdb11-cb9b-4cc4-8a0c-2be51837a591 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.800412] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.800866] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c05d4cd9-bbb9-4fef-a2fd-f65424755f20 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.804904] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 963.804904] env[62814]: value = "task-4294095" [ 963.804904] env[62814]: _type = "Task" [ 963.804904] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.811537] env[62814]: DEBUG oslo_vmware.api [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 963.811537] env[62814]: value = "task-4294096" [ 963.811537] env[62814]: _type = "Task" [ 963.811537] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.819677] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294095, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.835237] env[62814]: DEBUG oslo_vmware.api [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.966042] env[62814]: DEBUG nova.compute.utils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 963.970143] env[62814]: DEBUG nova.compute.manager [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 963.970143] env[62814]: DEBUG nova.network.neutron [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 964.058045] env[62814]: DEBUG nova.policy [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd05b1b390e0489d91f4f386f4c6fe41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65a886f76f404ce5a6d0841241d8f120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 964.059464] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquiring lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.059678] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquired lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 964.316184] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294095, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.325440] env[62814]: DEBUG oslo_vmware.api [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294096, 'name': PowerOffVM_Task, 'duration_secs': 0.223944} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.325738] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 964.325907] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 964.326259] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b5b2585-c8fb-419e-9e52-7b36549b9cba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.392989] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.393308] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.393498] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleting the datastore file [datastore2] 267c1ef2-ccc3-470e-943d-5fe9203c4dbc {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.393777] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36ef6456-8b9d-4963-a3d0-7144430c62ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.401534] env[62814]: DEBUG oslo_vmware.api [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 964.401534] env[62814]: value = "task-4294098" [ 964.401534] env[62814]: _type = "Task" [ 964.401534] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.411145] env[62814]: DEBUG oslo_vmware.api [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294098, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.471029] env[62814]: DEBUG nova.compute.manager [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 964.779981] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "9683894b-a300-4400-a1b9-db62478f42c5" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 964.779981] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 964.804432] env[62814]: DEBUG nova.network.neutron [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Successfully created port: 029acc9a-5528-4c96-a5a7-fa02e89a2bb9 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.818397] env[62814]: DEBUG oslo_vmware.api [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294095, 'name': RemoveSnapshot_Task, 'duration_secs': 0.580796} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.818732] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 964.819049] env[62814]: INFO nova.compute.manager [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Took 15.42 seconds to snapshot the instance on the hypervisor. [ 964.914460] env[62814]: DEBUG oslo_vmware.api [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294098, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17104} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.914460] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.914460] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 964.914460] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 964.914460] env[62814]: INFO nova.compute.manager [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Took 1.12 seconds to destroy the instance on the hypervisor. [ 964.914460] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 964.914460] env[62814]: DEBUG nova.compute.manager [-] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 964.914460] env[62814]: DEBUG nova.network.neutron [-] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 964.970777] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2bf9518-8fc2-40dd-a80b-809613e29b61 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.983071] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c58897-d965-4bee-903e-20c7ff5719b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.018946] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89a200cb-91c9-4b16-a296-10d749d555b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.028759] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5d7d26-576e-419f-a87d-58b474c5d562 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.044260] env[62814]: DEBUG nova.compute.provider_tree [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.099877] env[62814]: DEBUG oslo_concurrency.lockutils [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "6a592192-1b41-4be2-84a6-c3b76a4e5643" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.100250] env[62814]: DEBUG oslo_concurrency.lockutils [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.100524] env[62814]: DEBUG oslo_concurrency.lockutils [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "6a592192-1b41-4be2-84a6-c3b76a4e5643-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 965.100876] env[62814]: DEBUG oslo_concurrency.lockutils [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 965.102068] env[62814]: DEBUG oslo_concurrency.lockutils [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 965.103279] env[62814]: INFO nova.compute.manager [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Terminating instance [ 965.164751] env[62814]: DEBUG nova.compute.manager [req-2de7f91b-f5b2-4e84-9eb6-0c6c8cc94789 req-cd14ed09-2a5f-4615-8da8-d031acebfdff service nova] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Received event network-vif-deleted-3c36df9c-47da-4470-9e9d-87659604c923 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 965.206697] env[62814]: DEBUG nova.network.neutron [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 965.283368] env[62814]: DEBUG nova.compute.utils [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 965.394751] env[62814]: DEBUG nova.compute.manager [None req-ccd317c4-13cf-4b81-a3f5-2e4f73f97e65 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Found 2 images (rotation: 2) {{(pid=62814) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5017}} [ 965.483674] env[62814]: DEBUG nova.compute.manager [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 965.509110] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 965.509110] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.509281] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 965.509366] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.509513] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 965.509660] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 965.509939] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 965.510391] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 965.510594] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 965.510765] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 965.511699] env[62814]: DEBUG nova.virt.hardware [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 965.511904] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48b3d9f-6e7a-496f-aa29-3a5492c3fed7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.522600] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1178f8f-a5f4-4c3c-b981-12efc949d3c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.547788] env[62814]: DEBUG nova.scheduler.client.report [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 965.607786] env[62814]: DEBUG nova.compute.manager [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 965.608048] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 965.609607] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f55c62-7dfe-49cf-a125-9e5d5651a001 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.620184] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 965.620493] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d81d246-25bb-4ca4-b7a0-4577d2889953 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.634253] env[62814]: DEBUG oslo_vmware.api [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 965.634253] env[62814]: value = "task-4294099" [ 965.634253] env[62814]: _type = "Task" [ 965.634253] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.646802] env[62814]: DEBUG nova.compute.manager [req-b309ad53-ce43-44d1-a4bc-8ef90262bd0a req-4ee60107-f5f0-4630-ad08-8e5b284f391d service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Received event network-changed-1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 965.647011] env[62814]: DEBUG nova.compute.manager [req-b309ad53-ce43-44d1-a4bc-8ef90262bd0a req-4ee60107-f5f0-4630-ad08-8e5b284f391d service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Refreshing instance network info cache due to event network-changed-1865cb89-88d6-4e84-81ae-daece6adaa4d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 965.647250] env[62814]: DEBUG oslo_concurrency.lockutils [req-b309ad53-ce43-44d1-a4bc-8ef90262bd0a req-4ee60107-f5f0-4630-ad08-8e5b284f391d service nova] Acquiring lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.654540] env[62814]: DEBUG oslo_vmware.api [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4294099, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.788293] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.053982] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.596s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 966.054540] env[62814]: DEBUG nova.compute.manager [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 966.062511] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 25.061s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.062744] env[62814]: DEBUG nova.objects.instance [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 966.140139] env[62814]: DEBUG nova.network.neutron [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updating instance_info_cache with network_info: [{"id": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "address": "fa:16:3e:8e:bc:5f", "network": {"id": "ddc113f9-9fc4-4847-9181-82a3ee75ec5c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-787311776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6254d7e15b4c5ba267e8da03d1b82d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1865cb89-88", "ovs_interfaceid": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.149807] env[62814]: DEBUG oslo_vmware.api [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4294099, 'name': PowerOffVM_Task, 'duration_secs': 0.218643} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.150499] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 966.151132] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 966.151132] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9926865-f4ed-4769-9009-fffd5cfdcb78 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.214728] env[62814]: DEBUG nova.network.neutron [-] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.217660] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 966.217893] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 966.218118] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Deleting the datastore file [datastore2] 6a592192-1b41-4be2-84a6-c3b76a4e5643 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 966.218697] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81c7ee17-4f00-4a87-a226-b371166fd6de {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.226956] env[62814]: DEBUG oslo_vmware.api [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for the task: (returnval){ [ 966.226956] env[62814]: value = "task-4294101" [ 966.226956] env[62814]: _type = "Task" [ 966.226956] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.244248] env[62814]: DEBUG oslo_vmware.api [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4294101, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.568071] env[62814]: DEBUG nova.compute.utils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 966.572652] env[62814]: DEBUG nova.compute.manager [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 966.572652] env[62814]: DEBUG nova.network.neutron [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 966.642713] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Releasing lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 966.642993] env[62814]: DEBUG nova.compute.manager [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Inject network info {{(pid=62814) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7737}} [ 966.643304] env[62814]: DEBUG nova.compute.manager [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] network_info to inject: |[{"id": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "address": "fa:16:3e:8e:bc:5f", "network": {"id": "ddc113f9-9fc4-4847-9181-82a3ee75ec5c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-787311776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6254d7e15b4c5ba267e8da03d1b82d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1865cb89-88", "ovs_interfaceid": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7738}} [ 966.651438] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Reconfiguring VM instance to set the machine id {{(pid=62814) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 966.655233] env[62814]: DEBUG nova.policy [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64e84da705284d2dbf693a26ef184cd7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95c92336f9e746edba50b0b9e078b0dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 966.657169] env[62814]: DEBUG oslo_concurrency.lockutils [req-b309ad53-ce43-44d1-a4bc-8ef90262bd0a req-4ee60107-f5f0-4630-ad08-8e5b284f391d service nova] Acquired lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 966.657431] env[62814]: DEBUG nova.network.neutron [req-b309ad53-ce43-44d1-a4bc-8ef90262bd0a req-4ee60107-f5f0-4630-ad08-8e5b284f391d service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Refreshing network info cache for port 1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 966.659450] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ee18f54b-ff0b-485b-9359-7454e96db63d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.683089] env[62814]: DEBUG oslo_vmware.api [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 966.683089] env[62814]: value = "task-4294102" [ 966.683089] env[62814]: _type = "Task" [ 966.683089] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.699361] env[62814]: DEBUG oslo_vmware.api [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4294102, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.720196] env[62814]: INFO nova.compute.manager [-] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Took 1.81 seconds to deallocate network for instance. [ 966.739755] env[62814]: DEBUG oslo_vmware.api [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Task: {'id': task-4294101, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369667} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.740170] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 966.740453] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 966.740558] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 966.742298] env[62814]: INFO nova.compute.manager [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Took 1.13 seconds to destroy the instance on the hypervisor. [ 966.743060] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 966.743060] env[62814]: DEBUG nova.compute.manager [-] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 966.743060] env[62814]: DEBUG nova.network.neutron [-] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 966.902080] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "9683894b-a300-4400-a1b9-db62478f42c5" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 966.902545] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 966.902945] env[62814]: INFO nova.compute.manager [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Attaching volume f4d241a2-9ed8-46ae-af01-e7e67273236c to /dev/sdb [ 966.960732] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573caeed-d3cb-4759-8299-d9daf5603c82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.972663] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e4e34f-736a-40de-a755-1d0f7274824b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.991225] env[62814]: DEBUG nova.virt.block_device [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Updating existing volume attachment record: c2b95a49-b460-4332-bf01-3354e1e3d073 {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 967.012407] env[62814]: DEBUG nova.network.neutron [req-b309ad53-ce43-44d1-a4bc-8ef90262bd0a req-4ee60107-f5f0-4630-ad08-8e5b284f391d service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updated VIF entry in instance network info cache for port 1865cb89-88d6-4e84-81ae-daece6adaa4d. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 967.012646] env[62814]: DEBUG nova.network.neutron [req-b309ad53-ce43-44d1-a4bc-8ef90262bd0a req-4ee60107-f5f0-4630-ad08-8e5b284f391d service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updating instance_info_cache with network_info: [{"id": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "address": "fa:16:3e:8e:bc:5f", "network": {"id": "ddc113f9-9fc4-4847-9181-82a3ee75ec5c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-787311776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}, {"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6254d7e15b4c5ba267e8da03d1b82d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1865cb89-88", "ovs_interfaceid": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.076640] env[62814]: DEBUG nova.compute.manager [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 967.086056] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9fd4ee3c-1037-46ac-aa87-a05034a8907a tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.023s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.087891] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.165s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.088450] env[62814]: DEBUG nova.objects.instance [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lazy-loading 'resources' on Instance uuid 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.197598] env[62814]: DEBUG oslo_vmware.api [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4294102, 'name': ReconfigVM_Task, 'duration_secs': 0.189189} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.197598] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e0b37118-b649-4ff0-af17-b529a0e9eaca tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Reconfigured VM instance to set the machine id {{(pid=62814) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 967.228104] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 967.279541] env[62814]: DEBUG nova.objects.instance [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lazy-loading 'flavor' on Instance uuid d366a755-49b4-427b-8564-d8572a7fbbb7 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.480191] env[62814]: DEBUG nova.network.neutron [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Successfully updated port: 029acc9a-5528-4c96-a5a7-fa02e89a2bb9 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.508528] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52396f16-108b-5b64-8ecd-5e633af20030/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 967.511539] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4042cbe6-d42b-4704-8a35-779730c64edc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.516482] env[62814]: DEBUG oslo_concurrency.lockutils [req-b309ad53-ce43-44d1-a4bc-8ef90262bd0a req-4ee60107-f5f0-4630-ad08-8e5b284f391d service nova] Releasing lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 967.521662] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52396f16-108b-5b64-8ecd-5e633af20030/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 967.521901] env[62814]: ERROR oslo_vmware.rw_handles [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52396f16-108b-5b64-8ecd-5e633af20030/disk-0.vmdk due to incomplete transfer. [ 967.522117] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-602fa3e8-70da-4cf1-8480-00abfd20f388 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.525747] env[62814]: DEBUG nova.network.neutron [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Successfully created port: 260ddd77-4e00-4b1e-9299-fa71a7d23a57 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.530468] env[62814]: DEBUG oslo_vmware.rw_handles [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52396f16-108b-5b64-8ecd-5e633af20030/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 967.530673] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Uploaded image f876a8b9-10c0-49c4-8d83-20d0d8b591c4 to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 967.532412] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 967.532805] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-857e61c6-fa19-4bac-8632-111c5249db83 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.541746] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 967.541746] env[62814]: value = "task-4294104" [ 967.541746] env[62814]: _type = "Task" [ 967.541746] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.551583] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294104, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.599771] env[62814]: DEBUG nova.compute.manager [req-17b6395f-b846-4558-8c74-cf988fd88015 req-260a043e-b1bc-428f-b4be-a5da628ba681 service nova] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Received event network-vif-deleted-ef217261-6aa3-43f9-97b8-9c8d818e17f4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 967.600029] env[62814]: DEBUG nova.compute.manager [req-17b6395f-b846-4558-8c74-cf988fd88015 req-260a043e-b1bc-428f-b4be-a5da628ba681 service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Received event network-vif-plugged-029acc9a-5528-4c96-a5a7-fa02e89a2bb9 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 967.600188] env[62814]: DEBUG oslo_concurrency.lockutils [req-17b6395f-b846-4558-8c74-cf988fd88015 req-260a043e-b1bc-428f-b4be-a5da628ba681 service nova] Acquiring lock "40e2d845-0211-4c84-aef7-94014f999e1d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 967.600391] env[62814]: DEBUG oslo_concurrency.lockutils [req-17b6395f-b846-4558-8c74-cf988fd88015 req-260a043e-b1bc-428f-b4be-a5da628ba681 service nova] Lock "40e2d845-0211-4c84-aef7-94014f999e1d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 967.600553] env[62814]: DEBUG oslo_concurrency.lockutils [req-17b6395f-b846-4558-8c74-cf988fd88015 req-260a043e-b1bc-428f-b4be-a5da628ba681 service nova] Lock "40e2d845-0211-4c84-aef7-94014f999e1d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 967.600757] env[62814]: DEBUG nova.compute.manager [req-17b6395f-b846-4558-8c74-cf988fd88015 req-260a043e-b1bc-428f-b4be-a5da628ba681 service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] No waiting events found dispatching network-vif-plugged-029acc9a-5528-4c96-a5a7-fa02e89a2bb9 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 967.600927] env[62814]: WARNING nova.compute.manager [req-17b6395f-b846-4558-8c74-cf988fd88015 req-260a043e-b1bc-428f-b4be-a5da628ba681 service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Received unexpected event network-vif-plugged-029acc9a-5528-4c96-a5a7-fa02e89a2bb9 for instance with vm_state building and task_state spawning. [ 967.745394] env[62814]: DEBUG nova.compute.manager [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 967.745742] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed13b6a-eeec-4db1-80eb-1752904d51c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.787136] env[62814]: DEBUG oslo_concurrency.lockutils [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquiring lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.787136] env[62814]: DEBUG oslo_concurrency.lockutils [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquired lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 967.997919] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "refresh_cache-40e2d845-0211-4c84-aef7-94014f999e1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.997919] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "refresh_cache-40e2d845-0211-4c84-aef7-94014f999e1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 967.997919] env[62814]: DEBUG nova.network.neutron [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 967.997919] env[62814]: DEBUG nova.network.neutron [-] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.062386] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294104, 'name': Destroy_Task, 'duration_secs': 0.347667} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.062831] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Destroyed the VM [ 968.064349] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 968.064349] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-dca37311-d98b-4f5f-a7e0-caf2fa4755a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.073385] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 968.073385] env[62814]: value = "task-4294107" [ 968.073385] env[62814]: _type = "Task" [ 968.073385] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.093716] env[62814]: DEBUG nova.compute.manager [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 968.094146] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294107, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.134786] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 968.137071] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.137635] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 968.137635] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.137925] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 968.138199] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 968.138706] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 968.138993] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 968.139319] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 968.139652] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 968.139981] env[62814]: DEBUG nova.virt.hardware [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 968.141541] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-826f40f6-28a4-4b74-981c-da255f0ae3fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.145935] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65052f94-4abe-4f5c-8357-8c6d42b7268a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.158206] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eef2d159-7879-4f1e-9ced-81a02d6d2f47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.165469] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b201a4c-77f6-48a1-b095-3540ec6a518a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.211661] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7763613-d38d-4012-9ed0-23bb21ae56a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.220598] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec911f0b-dcb3-4318-b3fc-7f2a6e25259d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.234886] env[62814]: DEBUG nova.compute.provider_tree [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.257816] env[62814]: INFO nova.compute.manager [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] instance snapshotting [ 968.258453] env[62814]: DEBUG nova.objects.instance [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'flavor' on Instance uuid 9343301c-dfe9-41b0-b4a0-067af544d297 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.503011] env[62814]: INFO nova.compute.manager [-] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Took 1.76 seconds to deallocate network for instance. [ 968.562054] env[62814]: DEBUG nova.network.neutron [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 968.586731] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294107, 'name': RemoveSnapshot_Task} progress is 81%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.617460] env[62814]: DEBUG nova.network.neutron [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.737684] env[62814]: DEBUG nova.scheduler.client.report [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 968.764219] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ba48f0-515b-40bc-819e-45a7a5239ab2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.787857] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbbd7c7-cb9f-4c78-b039-5a0bd39dcd8c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.813402] env[62814]: DEBUG nova.network.neutron [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Updating instance_info_cache with network_info: [{"id": "029acc9a-5528-4c96-a5a7-fa02e89a2bb9", "address": "fa:16:3e:e8:ed:4e", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap029acc9a-55", "ovs_interfaceid": "029acc9a-5528-4c96-a5a7-fa02e89a2bb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.027094] env[62814]: DEBUG oslo_concurrency.lockutils [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 969.087586] env[62814]: DEBUG oslo_vmware.api [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294107, 'name': RemoveSnapshot_Task, 'duration_secs': 0.833233} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.087586] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 969.087586] env[62814]: INFO nova.compute.manager [None req-ec6e2f95-34c8-46c8-9330-957ea4daea6a tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Took 15.67 seconds to snapshot the instance on the hypervisor. [ 969.245030] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.156s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 969.249463] env[62814]: DEBUG oslo_concurrency.lockutils [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.149s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 969.249463] env[62814]: DEBUG nova.objects.instance [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lazy-loading 'resources' on Instance uuid 5c574786-64a4-4f07-a267-101ecaaa6938 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.278492] env[62814]: INFO nova.scheduler.client.report [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Deleted allocations for instance 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a [ 969.300349] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 969.304645] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d5fe6e52-98bd-411f-9cac-94f9ab3c4459 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.313153] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 969.313153] env[62814]: value = "task-4294108" [ 969.313153] env[62814]: _type = "Task" [ 969.313153] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.318193] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "refresh_cache-40e2d845-0211-4c84-aef7-94014f999e1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 969.319366] env[62814]: DEBUG nova.compute.manager [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Instance network_info: |[{"id": "029acc9a-5528-4c96-a5a7-fa02e89a2bb9", "address": "fa:16:3e:e8:ed:4e", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap029acc9a-55", "ovs_interfaceid": "029acc9a-5528-4c96-a5a7-fa02e89a2bb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 969.325181] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:ed:4e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7e0240aa-a694-48fc-a0f9-6f2d3e71aa12', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '029acc9a-5528-4c96-a5a7-fa02e89a2bb9', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.334781] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 969.335143] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294108, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.335792] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 969.336090] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57844771-1556-44d3-bd48-638864c3a158 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.364080] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.364080] env[62814]: value = "task-4294109" [ 969.364080] env[62814]: _type = "Task" [ 969.364080] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.378786] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294109, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.786920] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1eb6a7df-8212-4c0b-a9ea-41ddcd7e3ee3 tempest-ServersListShow296Test-2121141507 tempest-ServersListShow296Test-2121141507-project-member] Lock "0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.092s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 969.838826] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294108, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.874577] env[62814]: DEBUG nova.network.neutron [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updating instance_info_cache with network_info: [{"id": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "address": "fa:16:3e:8e:bc:5f", "network": {"id": "ddc113f9-9fc4-4847-9181-82a3ee75ec5c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-787311776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6254d7e15b4c5ba267e8da03d1b82d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1865cb89-88", "ovs_interfaceid": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.883922] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294109, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.939347] env[62814]: DEBUG nova.network.neutron [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Successfully updated port: 260ddd77-4e00-4b1e-9299-fa71a7d23a57 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 970.103134] env[62814]: DEBUG nova.compute.manager [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Received event network-changed-029acc9a-5528-4c96-a5a7-fa02e89a2bb9 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 970.103134] env[62814]: DEBUG nova.compute.manager [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Refreshing instance network info cache due to event network-changed-029acc9a-5528-4c96-a5a7-fa02e89a2bb9. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 970.103134] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] Acquiring lock "refresh_cache-40e2d845-0211-4c84-aef7-94014f999e1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.103134] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] Acquired lock "refresh_cache-40e2d845-0211-4c84-aef7-94014f999e1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 970.103134] env[62814]: DEBUG nova.network.neutron [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Refreshing network info cache for port 029acc9a-5528-4c96-a5a7-fa02e89a2bb9 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 970.260662] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80032589-687f-47a9-8a7a-48cfa3c09938 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.272785] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30bb3f1a-1223-4791-a9ec-eaa5c463edd6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.324967] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0dc1b75-0ca6-471f-8fb3-8309c6dea5cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.335912] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294108, 'name': CreateSnapshot_Task, 'duration_secs': 0.552513} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.336567] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 970.337402] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66961fbb-d428-4679-88f6-1c001c29b1b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.340760] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d29819c-2045-43b7-bf02-a654bd386627 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.366411] env[62814]: DEBUG nova.compute.provider_tree [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 970.376850] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294109, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.389228] env[62814]: DEBUG oslo_concurrency.lockutils [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Releasing lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 970.389484] env[62814]: DEBUG nova.compute.manager [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Inject network info {{(pid=62814) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7737}} [ 970.389719] env[62814]: DEBUG nova.compute.manager [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] network_info to inject: |[{"id": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "address": "fa:16:3e:8e:bc:5f", "network": {"id": "ddc113f9-9fc4-4847-9181-82a3ee75ec5c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-787311776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6254d7e15b4c5ba267e8da03d1b82d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1865cb89-88", "ovs_interfaceid": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7738}} [ 970.394783] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Reconfiguring VM instance to set the machine id {{(pid=62814) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 970.395471] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae8056ca-1b28-49a4-b310-ab17538469f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.415122] env[62814]: DEBUG oslo_vmware.api [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 970.415122] env[62814]: value = "task-4294111" [ 970.415122] env[62814]: _type = "Task" [ 970.415122] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.423191] env[62814]: DEBUG oslo_vmware.api [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4294111, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.448122] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-b5ddd05f-2027-4edf-84c5-0d2f537a95e4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.448287] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-b5ddd05f-2027-4edf-84c5-0d2f537a95e4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 970.448442] env[62814]: DEBUG nova.network.neutron [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.877639] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 970.886076] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-78ba0f0d-d683-48b6-84f7-fe44fb3fdc4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.889946] env[62814]: DEBUG nova.network.neutron [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Updated VIF entry in instance network info cache for port 029acc9a-5528-4c96-a5a7-fa02e89a2bb9. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.890570] env[62814]: DEBUG nova.network.neutron [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Updating instance_info_cache with network_info: [{"id": "029acc9a-5528-4c96-a5a7-fa02e89a2bb9", "address": "fa:16:3e:e8:ed:4e", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap029acc9a-55", "ovs_interfaceid": "029acc9a-5528-4c96-a5a7-fa02e89a2bb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.903524] env[62814]: DEBUG oslo_concurrency.lockutils [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 970.903524] env[62814]: DEBUG oslo_concurrency.lockutils [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 970.903524] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294109, 'name': CreateVM_Task, 'duration_secs': 1.342724} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.904396] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 970.904755] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 970.904755] env[62814]: value = "task-4294112" [ 970.904755] env[62814]: _type = "Task" [ 970.904755] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.905857] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.906026] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 970.906364] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 970.906850] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-482754dc-805f-49b1-9bc7-ba997c9f6d70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.915802] env[62814]: ERROR nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [req-93fe6961-e05e-4ac5-a32d-63d22f175d19] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-93fe6961-e05e-4ac5-a32d-63d22f175d19"}]} [ 970.928246] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 970.928246] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52044347-6b2b-2133-4d6a-2f27806d5ac2" [ 970.928246] env[62814]: _type = "Task" [ 970.928246] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.928246] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294112, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.936032] env[62814]: DEBUG oslo_vmware.api [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4294111, 'name': ReconfigVM_Task, 'duration_secs': 0.156669} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.936643] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-06d27da3-b72d-45a9-b41f-6937b33a33a2 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Reconfigured VM instance to set the machine id {{(pid=62814) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 970.944284] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52044347-6b2b-2133-4d6a-2f27806d5ac2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.949199] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 970.964994] env[62814]: DEBUG nova.compute.manager [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 970.966355] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a51476-f19e-413b-a784-99ed2fbdf9fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.980985] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 970.984353] env[62814]: DEBUG nova.compute.provider_tree [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 970.999705] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 971.014252] env[62814]: DEBUG nova.network.neutron [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 971.042636] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 971.191741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquiring lock "d366a755-49b4-427b-8564-d8572a7fbbb7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.191741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lock "d366a755-49b4-427b-8564-d8572a7fbbb7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 971.191741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquiring lock "d366a755-49b4-427b-8564-d8572a7fbbb7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 971.192084] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lock "d366a755-49b4-427b-8564-d8572a7fbbb7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 971.192738] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lock "d366a755-49b4-427b-8564-d8572a7fbbb7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 971.197157] env[62814]: INFO nova.compute.manager [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Terminating instance [ 971.400110] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] Releasing lock "refresh_cache-40e2d845-0211-4c84-aef7-94014f999e1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.401977] env[62814]: DEBUG nova.compute.manager [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Received event network-vif-deleted-827a94b4-5864-4060-bfb5-d0e9d2281332 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 971.401977] env[62814]: DEBUG nova.compute.manager [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Received event network-changed-1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 971.401977] env[62814]: DEBUG nova.compute.manager [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Refreshing instance network info cache due to event network-changed-1865cb89-88d6-4e84-81ae-daece6adaa4d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 971.401977] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] Acquiring lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.401977] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] Acquired lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 971.401977] env[62814]: DEBUG nova.network.neutron [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Refreshing network info cache for port 1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 971.408669] env[62814]: INFO nova.compute.manager [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Detaching volume 252ca9ac-8b1f-4449-8f58-9fce53fe27ec [ 971.417442] env[62814]: DEBUG nova.network.neutron [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Updating instance_info_cache with network_info: [{"id": "260ddd77-4e00-4b1e-9299-fa71a7d23a57", "address": "fa:16:3e:be:4b:c0", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap260ddd77-4e", "ovs_interfaceid": "260ddd77-4e00-4b1e-9299-fa71a7d23a57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.423027] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294112, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.442872] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52044347-6b2b-2133-4d6a-2f27806d5ac2, 'name': SearchDatastore_Task, 'duration_secs': 0.02315} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.445729] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.445977] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 971.446230] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.446381] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 971.446563] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.450555] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a97e73fd-2a85-4af2-a056-acc24b5af026 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.456965] env[62814]: INFO nova.virt.block_device [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Attempting to driver detach volume 252ca9ac-8b1f-4449-8f58-9fce53fe27ec from mountpoint /dev/sdb [ 971.458253] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 971.458253] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845734', 'volume_id': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'name': 'volume-252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3e4479a0-8edd-4b37-8cc9-2c91275b88ee', 'attached_at': '', 'detached_at': '', 'volume_id': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'serial': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 971.458591] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6ea714-ddc3-4328-a768-62db13e2249f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.465928] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.465928] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 971.484669] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e311a1d4-9749-457a-a10d-8d162b30dbab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.488372] env[62814]: INFO nova.compute.manager [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] instance snapshotting [ 971.490977] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac1d119a-68cf-45d1-a497-0f7f5e3daf44 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.496920] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef676d1-c693-438d-8b8b-fa906105c179 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.501969] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 971.501969] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5291957c-b22a-5385-367e-af4034553783" [ 971.501969] env[62814]: _type = "Task" [ 971.501969] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.525288] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697c5843-2166-4bf4-b8c1-2b0b3bd3044a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.531140] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42dfe010-1171-4c9c-b2a2-eafcfa2f92d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.539497] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5291957c-b22a-5385-367e-af4034553783, 'name': SearchDatastore_Task, 'duration_secs': 0.019998} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.539956] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7b51c3a-f6e5-45cd-94dc-26f199f72d10 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.570188] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-611e2e72-e1a0-4c83-9aff-a36d48da1457 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.574544] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 971.574544] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520cc632-7867-0593-31f4-c9d55379fde8" [ 971.574544] env[62814]: _type = "Task" [ 971.574544] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.589267] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] The volume has not been displaced from its original location: [datastore1] volume-252ca9ac-8b1f-4449-8f58-9fce53fe27ec/volume-252ca9ac-8b1f-4449-8f58-9fce53fe27ec.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 971.594846] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Reconfiguring VM instance instance-0000002b to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 971.597186] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1218f461-9324-4bba-a536-c4d26717d7e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.610240] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d969a2b-ccff-4df4-8690-c22cd21d4ae5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.617152] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520cc632-7867-0593-31f4-c9d55379fde8, 'name': SearchDatastore_Task, 'duration_secs': 0.014559} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.618131] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.618421] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 40e2d845-0211-4c84-aef7-94014f999e1d/40e2d845-0211-4c84-aef7-94014f999e1d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 971.619383] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec338e0c-69bf-40e1-83a6-2212dee25ed2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.625754] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5316700-c27e-48fb-b084-8348d612a1f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.629093] env[62814]: DEBUG oslo_vmware.api [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 971.629093] env[62814]: value = "task-4294113" [ 971.629093] env[62814]: _type = "Task" [ 971.629093] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.666488] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 971.666488] env[62814]: value = "task-4294114" [ 971.666488] env[62814]: _type = "Task" [ 971.666488] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.671603] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da304bf-7add-4665-bdc9-fe8b4a7070a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.681399] env[62814]: DEBUG oslo_vmware.api [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294113, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.693018] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-275194c0-416c-40e9-81cd-bb9f346bf169 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.697431] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294114, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.710032] env[62814]: DEBUG nova.compute.manager [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 971.710032] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 971.710581] env[62814]: DEBUG nova.compute.provider_tree [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 971.714344] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b14eef0-d83b-4043-ba21-053f7401de69 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.721834] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 971.722043] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0bdef0a4-8800-4e98-98a9-234abb024ae7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.731944] env[62814]: DEBUG oslo_vmware.api [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 971.731944] env[62814]: value = "task-4294115" [ 971.731944] env[62814]: _type = "Task" [ 971.731944] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.744376] env[62814]: DEBUG oslo_vmware.api [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4294115, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.926784] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-b5ddd05f-2027-4edf-84c5-0d2f537a95e4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 971.930799] env[62814]: DEBUG nova.compute.manager [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Instance network_info: |[{"id": "260ddd77-4e00-4b1e-9299-fa71a7d23a57", "address": "fa:16:3e:be:4b:c0", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap260ddd77-4e", "ovs_interfaceid": "260ddd77-4e00-4b1e-9299-fa71a7d23a57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 971.930799] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294112, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.930799] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:4b:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '605f83bd-808c-4b54-922e-54b14690987a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '260ddd77-4e00-4b1e-9299-fa71a7d23a57', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 971.937307] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Creating folder: Project (95c92336f9e746edba50b0b9e078b0dd). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 971.937646] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b284c45-07ea-4280-b459-d6f8923a28af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.953935] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Created folder: Project (95c92336f9e746edba50b0b9e078b0dd) in parent group-v845547. [ 971.954161] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Creating folder: Instances. Parent ref: group-v845763. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 971.954484] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-70ef251f-7d3a-4320-8bd9-9159ef83352f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.970401] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Created folder: Instances in parent group-v845763. [ 971.970766] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 971.970994] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 971.971258] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80407b0d-b7e7-481f-9427-4eed8c6fefa1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.000679] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 972.000679] env[62814]: value = "task-4294118" [ 972.000679] env[62814]: _type = "Task" [ 972.000679] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.012805] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294118, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.073876] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 972.074959] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845759', 'volume_id': 'f4d241a2-9ed8-46ae-af01-e7e67273236c', 'name': 'volume-f4d241a2-9ed8-46ae-af01-e7e67273236c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9683894b-a300-4400-a1b9-db62478f42c5', 'attached_at': '', 'detached_at': '', 'volume_id': 'f4d241a2-9ed8-46ae-af01-e7e67273236c', 'serial': 'f4d241a2-9ed8-46ae-af01-e7e67273236c'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 972.076468] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 972.077634] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58dc7296-8880-450f-b298-c6e9707f10d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.080832] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-061202db-6bd4-4251-910d-15dcb65c3692 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.114328] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 972.114328] env[62814]: value = "task-4294119" [ 972.114328] env[62814]: _type = "Task" [ 972.114328] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.115322] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70bbd7c6-ca13-4de6-ba1c-ff5cce1a4d50 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.157327] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] volume-f4d241a2-9ed8-46ae-af01-e7e67273236c/volume-f4d241a2-9ed8-46ae-af01-e7e67273236c.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.161166] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-062d233b-17fa-43b2-bc2c-7fd8d5b7c3e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.187653] env[62814]: DEBUG oslo_vmware.api [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294113, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.197468] env[62814]: DEBUG oslo_vmware.api [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 972.197468] env[62814]: value = "task-4294120" [ 972.197468] env[62814]: _type = "Task" [ 972.197468] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.197754] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294114, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.208504] env[62814]: DEBUG oslo_vmware.api [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294120, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.247917] env[62814]: DEBUG oslo_vmware.api [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4294115, 'name': PowerOffVM_Task, 'duration_secs': 0.28709} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.249085] env[62814]: ERROR nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [req-e4ce651a-5007-48e5-8e0f-0f5aa4db6b20] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e4ce651a-5007-48e5-8e0f-0f5aa4db6b20"}]} [ 972.249747] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 972.249919] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 972.253627] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe15e00e-901c-46e6-9daa-371b3b8994f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.278126] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 972.305965] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 972.306224] env[62814]: DEBUG nova.compute.provider_tree [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.330430] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 972.341176] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 972.341176] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 972.342168] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Deleting the datastore file [datastore2] d366a755-49b4-427b-8564-d8572a7fbbb7 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.342537] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c25d857c-cc00-4747-b30b-31105dd29f4a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.352850] env[62814]: DEBUG oslo_vmware.api [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for the task: (returnval){ [ 972.352850] env[62814]: value = "task-4294122" [ 972.352850] env[62814]: _type = "Task" [ 972.352850] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.365821] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 972.374642] env[62814]: DEBUG oslo_vmware.api [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4294122, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.392139] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquiring lock "a68901a9-9a9b-4127-bca9-64f98dfb151f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 972.392139] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Lock "a68901a9-9a9b-4127-bca9-64f98dfb151f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.431812] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294112, 'name': CloneVM_Task} progress is 95%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.441872] env[62814]: DEBUG nova.compute.manager [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Received event network-vif-plugged-260ddd77-4e00-4b1e-9299-fa71a7d23a57 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 972.442087] env[62814]: DEBUG oslo_concurrency.lockutils [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] Acquiring lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 972.442394] env[62814]: DEBUG oslo_concurrency.lockutils [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] Lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 972.442565] env[62814]: DEBUG oslo_concurrency.lockutils [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] Lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 972.442727] env[62814]: DEBUG nova.compute.manager [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] No waiting events found dispatching network-vif-plugged-260ddd77-4e00-4b1e-9299-fa71a7d23a57 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 972.442888] env[62814]: WARNING nova.compute.manager [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Received unexpected event network-vif-plugged-260ddd77-4e00-4b1e-9299-fa71a7d23a57 for instance with vm_state building and task_state spawning. [ 972.443063] env[62814]: DEBUG nova.compute.manager [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Received event network-changed-260ddd77-4e00-4b1e-9299-fa71a7d23a57 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 972.443217] env[62814]: DEBUG nova.compute.manager [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Refreshing instance network info cache due to event network-changed-260ddd77-4e00-4b1e-9299-fa71a7d23a57. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 972.443411] env[62814]: DEBUG oslo_concurrency.lockutils [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] Acquiring lock "refresh_cache-b5ddd05f-2027-4edf-84c5-0d2f537a95e4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.443657] env[62814]: DEBUG oslo_concurrency.lockutils [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] Acquired lock "refresh_cache-b5ddd05f-2027-4edf-84c5-0d2f537a95e4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 972.443698] env[62814]: DEBUG nova.network.neutron [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Refreshing network info cache for port 260ddd77-4e00-4b1e-9299-fa71a7d23a57 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 972.462953] env[62814]: DEBUG nova.network.neutron [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updated VIF entry in instance network info cache for port 1865cb89-88d6-4e84-81ae-daece6adaa4d. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 972.463612] env[62814]: DEBUG nova.network.neutron [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updating instance_info_cache with network_info: [{"id": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "address": "fa:16:3e:8e:bc:5f", "network": {"id": "ddc113f9-9fc4-4847-9181-82a3ee75ec5c", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-787311776-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fd6254d7e15b4c5ba267e8da03d1b82d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c47e98ff-83cf-48d2-bf91-2931c7386b6a", "external-id": "nsx-vlan-transportzone-992", "segmentation_id": 992, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1865cb89-88", "ovs_interfaceid": "1865cb89-88d6-4e84-81ae-daece6adaa4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.513747] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294118, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.629803] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294119, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.657049] env[62814]: DEBUG oslo_vmware.api [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294113, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.694024] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294114, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.658844} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.694024] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 40e2d845-0211-4c84-aef7-94014f999e1d/40e2d845-0211-4c84-aef7-94014f999e1d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 972.694024] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 972.694276] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5905d391-50e5-4561-9e8a-b4b45fd2795f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.707049] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 972.707049] env[62814]: value = "task-4294123" [ 972.707049] env[62814]: _type = "Task" [ 972.707049] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.716913] env[62814]: DEBUG oslo_vmware.api [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294120, 'name': ReconfigVM_Task, 'duration_secs': 0.501184} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.720304] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Reconfigured VM instance instance-00000035 to attach disk [datastore1] volume-f4d241a2-9ed8-46ae-af01-e7e67273236c/volume-f4d241a2-9ed8-46ae-af01-e7e67273236c.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.726337] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294123, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.728427] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-df3e66a2-4988-4200-a721-900533811dea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.748139] env[62814]: DEBUG oslo_vmware.api [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 972.748139] env[62814]: value = "task-4294124" [ 972.748139] env[62814]: _type = "Task" [ 972.748139] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.763041] env[62814]: DEBUG oslo_vmware.api [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294124, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.868774] env[62814]: DEBUG oslo_vmware.api [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Task: {'id': task-4294122, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.219738} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.869145] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.869365] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.869560] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.869732] env[62814]: INFO nova.compute.manager [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 972.869993] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 972.870252] env[62814]: DEBUG nova.compute.manager [-] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 972.870352] env[62814]: DEBUG nova.network.neutron [-] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 972.873533] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab266b9c-6cc6-4940-865c-3d03afde34ee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.886107] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727f5c50-6b84-494d-a90d-16a97a3bfb20 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.921929] env[62814]: DEBUG nova.compute.manager [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 972.929259] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5670c9-18e5-45f7-9c3f-fb6c83e3140b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.939656] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7337b3ca-a0dc-4514-86e9-5d97a5b656e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.954607] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294112, 'name': CloneVM_Task} progress is 95%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.966497] env[62814]: DEBUG nova.compute.provider_tree [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 972.967171] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a433bd9-d626-4f6b-927e-39bc8a5e126d req-830f8c5d-a373-4c34-9d83-113fec3c01cb service nova] Releasing lock "refresh_cache-d366a755-49b4-427b-8564-d8572a7fbbb7" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 973.016927] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294118, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.137304] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294119, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.173805] env[62814]: DEBUG oslo_vmware.api [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294113, 'name': ReconfigVM_Task, 'duration_secs': 1.367705} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.173805] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Reconfigured VM instance instance-0000002b to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 973.185788] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e546fe32-2740-4a09-a252-43fddf16513b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.236164] env[62814]: DEBUG oslo_vmware.api [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 973.236164] env[62814]: value = "task-4294125" [ 973.236164] env[62814]: _type = "Task" [ 973.236164] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.257346] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294123, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.143692} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.271722] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 973.278791] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af46e3bb-f76e-4948-b85f-de2795f1b18c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.282422] env[62814]: DEBUG oslo_vmware.api [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294125, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.324891] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 40e2d845-0211-4c84-aef7-94014f999e1d/40e2d845-0211-4c84-aef7-94014f999e1d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.325321] env[62814]: DEBUG oslo_vmware.api [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294124, 'name': ReconfigVM_Task, 'duration_secs': 0.311935} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.325622] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5333774a-c3e6-449e-ba5b-0443840f9674 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.343418] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845759', 'volume_id': 'f4d241a2-9ed8-46ae-af01-e7e67273236c', 'name': 'volume-f4d241a2-9ed8-46ae-af01-e7e67273236c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9683894b-a300-4400-a1b9-db62478f42c5', 'attached_at': '', 'detached_at': '', 'volume_id': 'f4d241a2-9ed8-46ae-af01-e7e67273236c', 'serial': 'f4d241a2-9ed8-46ae-af01-e7e67273236c'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 973.358679] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 973.358679] env[62814]: value = "task-4294126" [ 973.358679] env[62814]: _type = "Task" [ 973.358679] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.364252] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294126, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.450472] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294112, 'name': CloneVM_Task, 'duration_secs': 2.529999} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.452583] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Created linked-clone VM from snapshot [ 973.452583] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8290669-6fb4-4c7f-9600-ae23503f3f51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.464219] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Uploading image e7d28230-ebf8-44c9-b3b7-03c29d4f076a {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 973.499243] env[62814]: ERROR nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] [req-f2e52ca7-f121-4896-82b5-89f9e3582ee9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f2e52ca7-f121-4896-82b5-89f9e3582ee9"}]} [ 973.501963] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 973.522658] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294118, 'name': CreateVM_Task, 'duration_secs': 1.303302} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.522658] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 973.523272] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.523311] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 973.523664] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 973.527142] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89ffad38-d35a-4a74-93a5-526f82a1d8b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.528861] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 973.548686] env[62814]: DEBUG oslo_vmware.rw_handles [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 973.548686] env[62814]: value = "vm-845762" [ 973.548686] env[62814]: _type = "VirtualMachine" [ 973.548686] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 973.548686] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-52c56731-6651-4cd6-9453-0bd0e708d53e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.548686] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 973.548686] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52edbb21-df58-e1fd-86c7-e39f5280a92e" [ 973.548686] env[62814]: _type = "Task" [ 973.548686] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.562759] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52edbb21-df58-e1fd-86c7-e39f5280a92e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.563139] env[62814]: DEBUG oslo_vmware.rw_handles [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lease: (returnval){ [ 973.563139] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5249e87a-9dc2-f8d2-b7a8-93c7b56c3475" [ 973.563139] env[62814]: _type = "HttpNfcLease" [ 973.563139] env[62814]: } obtained for exporting VM: (result){ [ 973.563139] env[62814]: value = "vm-845762" [ 973.563139] env[62814]: _type = "VirtualMachine" [ 973.563139] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 973.563394] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the lease: (returnval){ [ 973.563394] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5249e87a-9dc2-f8d2-b7a8-93c7b56c3475" [ 973.563394] env[62814]: _type = "HttpNfcLease" [ 973.563394] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 973.567271] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 973.567480] env[62814]: DEBUG nova.compute.provider_tree [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 973.576657] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 973.576657] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5249e87a-9dc2-f8d2-b7a8-93c7b56c3475" [ 973.576657] env[62814]: _type = "HttpNfcLease" [ 973.576657] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 973.602157] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 973.632449] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294119, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.647337] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 973.787204] env[62814]: DEBUG oslo_vmware.api [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294125, 'name': ReconfigVM_Task, 'duration_secs': 0.169} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.787204] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845734', 'volume_id': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'name': 'volume-252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '3e4479a0-8edd-4b37-8cc9-2c91275b88ee', 'attached_at': '', 'detached_at': '', 'volume_id': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec', 'serial': '252ca9ac-8b1f-4449-8f58-9fce53fe27ec'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 973.870560] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294126, 'name': ReconfigVM_Task, 'duration_secs': 0.474454} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.878367] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 40e2d845-0211-4c84-aef7-94014f999e1d/40e2d845-0211-4c84-aef7-94014f999e1d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 973.879261] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c047914c-6584-4431-8e0e-fd27aa9c6b77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.887698] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 973.887698] env[62814]: value = "task-4294128" [ 973.887698] env[62814]: _type = "Task" [ 973.887698] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.898154] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294128, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.966142] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "e727eeb1-c5d0-4591-80bb-31746bf976a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 973.966432] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "e727eeb1-c5d0-4591-80bb-31746bf976a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 974.060198] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52edbb21-df58-e1fd-86c7-e39f5280a92e, 'name': SearchDatastore_Task, 'duration_secs': 0.025609} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.064560] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 974.064560] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 974.064560] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.064560] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 974.064560] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 974.064560] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cc40babd-19b2-463b-89e7-9e5a0cc74310 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.074279] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 974.074279] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5249e87a-9dc2-f8d2-b7a8-93c7b56c3475" [ 974.074279] env[62814]: _type = "HttpNfcLease" [ 974.074279] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 974.074631] env[62814]: DEBUG oslo_vmware.rw_handles [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 974.074631] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5249e87a-9dc2-f8d2-b7a8-93c7b56c3475" [ 974.074631] env[62814]: _type = "HttpNfcLease" [ 974.074631] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 974.075361] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c23579-bce2-49a4-84fb-d86fe713b9ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.081496] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 974.081683] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 974.082895] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-813ce84e-cee5-4546-9b83-5392eab48aee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.090298] env[62814]: DEBUG oslo_vmware.rw_handles [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5282c603-1bc8-2fe0-87ac-d4abf2f60ef2/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 974.090298] env[62814]: DEBUG oslo_vmware.rw_handles [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5282c603-1bc8-2fe0-87ac-d4abf2f60ef2/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 974.096863] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 974.096863] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528e0bdd-0120-3f51-045c-3982588f4d69" [ 974.096863] env[62814]: _type = "Task" [ 974.096863] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.170494] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528e0bdd-0120-3f51-045c-3982588f4d69, 'name': SearchDatastore_Task, 'duration_secs': 0.021601} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.170494] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22481587-c16e-45dd-b91f-e45231512173 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.177068] env[62814]: DEBUG nova.network.neutron [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Updated VIF entry in instance network info cache for port 260ddd77-4e00-4b1e-9299-fa71a7d23a57. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.177443] env[62814]: DEBUG nova.network.neutron [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Updating instance_info_cache with network_info: [{"id": "260ddd77-4e00-4b1e-9299-fa71a7d23a57", "address": "fa:16:3e:be:4b:c0", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap260ddd77-4e", "ovs_interfaceid": "260ddd77-4e00-4b1e-9299-fa71a7d23a57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.178967] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294119, 'name': CreateSnapshot_Task, 'duration_secs': 1.549029} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.179617] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 974.180368] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ed4098e-7141-41a6-b238-86c828fb6991 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.188225] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 974.188225] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a0cebd-f7ca-6e34-d306-0f987b1338c1" [ 974.188225] env[62814]: _type = "Task" [ 974.188225] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.200433] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1a05171a-07bb-4419-9681-8276671a832c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.206386] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a0cebd-f7ca-6e34-d306-0f987b1338c1, 'name': SearchDatastore_Task, 'duration_secs': 0.011235} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.209459] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 974.209738] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b5ddd05f-2027-4edf-84c5-0d2f537a95e4/b5ddd05f-2027-4edf-84c5-0d2f537a95e4.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 974.210190] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9db718f6-1515-4ba0-b524-bc406f720578 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.217752] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 974.217752] env[62814]: value = "task-4294129" [ 974.217752] env[62814]: _type = "Task" [ 974.217752] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.226487] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294129, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.350681] env[62814]: DEBUG nova.objects.instance [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lazy-loading 'flavor' on Instance uuid 3e4479a0-8edd-4b37-8cc9-2c91275b88ee {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.402641] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294128, 'name': Rename_Task, 'duration_secs': 0.246873} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.402641] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 974.402641] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0e1709a-1454-494d-872d-f1c598f063dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.414392] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 974.414392] env[62814]: value = "task-4294130" [ 974.414392] env[62814]: _type = "Task" [ 974.414392] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.423738] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294130, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.439183] env[62814]: DEBUG nova.objects.instance [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'flavor' on Instance uuid 9683894b-a300-4400-a1b9-db62478f42c5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.455769] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374a2941-6f38-4e5b-b31d-8c8da82c6714 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.472989] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a6ac47-e8a0-4780-93ba-4962c353b0de {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.481527] env[62814]: DEBUG nova.compute.manager [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 974.539639] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95feeea4-94bf-459f-812d-7f35a121d05b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.553867] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbc1dc09-13ff-4b08-be67-76d6fd4bd4ad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.577209] env[62814]: DEBUG nova.compute.provider_tree [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 974.687819] env[62814]: DEBUG oslo_concurrency.lockutils [req-c07b733a-a3b8-4c7a-9bc5-31a9b68dca54 req-de9d340d-1bbf-4eaa-8f77-ff611a730d76 service nova] Releasing lock "refresh_cache-b5ddd05f-2027-4edf-84c5-0d2f537a95e4" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 974.709989] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 974.710498] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2bb90be3-1332-4da8-a629-2580449515eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.723541] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 974.723541] env[62814]: value = "task-4294131" [ 974.723541] env[62814]: _type = "Task" [ 974.723541] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.730715] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294129, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499076} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.731454] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b5ddd05f-2027-4edf-84c5-0d2f537a95e4/b5ddd05f-2027-4edf-84c5-0d2f537a95e4.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.731710] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.731977] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-31d31a2f-5bb2-442c-b5e0-0fb0d727bdef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.737808] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294131, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.746238] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 974.746238] env[62814]: value = "task-4294132" [ 974.746238] env[62814]: _type = "Task" [ 974.746238] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.759348] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294132, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.930236] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294130, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.944745] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d96159d3-d843-47da-a04f-b0d1df94eec8 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.042s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.062173] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 975.143721] env[62814]: DEBUG nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 109 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 975.144180] env[62814]: DEBUG nova.compute.provider_tree [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 109 to 110 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 975.144258] env[62814]: DEBUG nova.compute.provider_tree [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 975.240319] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294131, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.262450] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294132, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08539} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.262450] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 975.262450] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de97196-c70e-4652-b8ff-8ef5895fe728 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.292494] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] b5ddd05f-2027-4edf-84c5-0d2f537a95e4/b5ddd05f-2027-4edf-84c5-0d2f537a95e4.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.292494] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82687385-4172-4a6c-8f85-3d5e464ed1fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.314842] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 975.314842] env[62814]: value = "task-4294133" [ 975.314842] env[62814]: _type = "Task" [ 975.314842] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.325774] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294133, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.362147] env[62814]: DEBUG oslo_concurrency.lockutils [None req-95514846-13ef-4f73-b464-8783cbc82c7d tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.461s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.398570] env[62814]: DEBUG nova.compute.manager [req-f357bfcd-1096-451c-a531-0affde7fa91f req-1e02322c-d1b8-4fb9-9f2b-b2f082b6e23b service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Received event network-vif-deleted-1865cb89-88d6-4e84-81ae-daece6adaa4d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 975.398570] env[62814]: INFO nova.compute.manager [req-f357bfcd-1096-451c-a531-0affde7fa91f req-1e02322c-d1b8-4fb9-9f2b-b2f082b6e23b service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Neutron deleted interface 1865cb89-88d6-4e84-81ae-daece6adaa4d; detaching it from the instance and deleting it from the info cache [ 975.398570] env[62814]: DEBUG nova.network.neutron [req-f357bfcd-1096-451c-a531-0affde7fa91f req-1e02322c-d1b8-4fb9-9f2b-b2f082b6e23b service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.425568] env[62814]: DEBUG oslo_vmware.api [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294130, 'name': PowerOnVM_Task, 'duration_secs': 0.845833} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.425855] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.426079] env[62814]: INFO nova.compute.manager [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Took 9.94 seconds to spawn the instance on the hypervisor. [ 975.426268] env[62814]: DEBUG nova.compute.manager [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 975.427055] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8131194d-7254-4b20-8f8e-4d89d21043f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.499996] env[62814]: DEBUG nova.network.neutron [-] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.650000] env[62814]: DEBUG oslo_concurrency.lockutils [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 6.403s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 975.652573] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.045s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 975.652748] env[62814]: DEBUG nova.objects.instance [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lazy-loading 'resources' on Instance uuid 9be62576-5a05-473f-befd-b33f5fde9185 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.750408] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294131, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.825260] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294133, 'name': ReconfigVM_Task, 'duration_secs': 0.501661} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.825698] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Reconfigured VM instance instance-0000004b to attach disk [datastore2] b5ddd05f-2027-4edf-84c5-0d2f537a95e4/b5ddd05f-2027-4edf-84c5-0d2f537a95e4.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.826409] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4547786-d3af-4f56-a253-78d11062c40f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.835596] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 975.835596] env[62814]: value = "task-4294134" [ 975.835596] env[62814]: _type = "Task" [ 975.835596] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.845084] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294134, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.902846] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83688894-1cf3-4899-b0d5-c0e993af4008 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.914957] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d844a9a-1015-4616-b8c8-6f8cea0841a3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.956297] env[62814]: DEBUG nova.compute.manager [req-f357bfcd-1096-451c-a531-0affde7fa91f req-1e02322c-d1b8-4fb9-9f2b-b2f082b6e23b service nova] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Detach interface failed, port_id=1865cb89-88d6-4e84-81ae-daece6adaa4d, reason: Instance d366a755-49b4-427b-8564-d8572a7fbbb7 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 976.004251] env[62814]: INFO nova.compute.manager [-] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Took 3.13 seconds to deallocate network for instance. [ 976.165571] env[62814]: INFO nova.scheduler.client.report [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Deleted allocations for instance 5c574786-64a4-4f07-a267-101ecaaa6938 [ 976.243120] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294131, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.354646] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294134, 'name': Rename_Task, 'duration_secs': 0.159837} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.354646] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 976.354646] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-683f1836-2fe6-435b-8670-b264d889494e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.364594] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 976.364594] env[62814]: value = "task-4294135" [ 976.364594] env[62814]: _type = "Task" [ 976.364594] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.382594] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294135, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.455107] env[62814]: INFO nova.compute.manager [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Took 37.90 seconds to build instance. [ 976.512707] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 976.617614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "9683894b-a300-4400-a1b9-db62478f42c5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 976.617614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 976.617614] env[62814]: DEBUG nova.compute.manager [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 976.618707] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a7bdd7-749f-4e1f-878f-df4299747827 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.640817] env[62814]: DEBUG nova.compute.manager [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 976.640817] env[62814]: DEBUG nova.objects.instance [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'flavor' on Instance uuid 9683894b-a300-4400-a1b9-db62478f42c5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.745702] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294131, 'name': CloneVM_Task, 'duration_secs': 1.794614} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.746772] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Created linked-clone VM from snapshot [ 976.746882] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0099149f-8943-4146-a89d-82c4a002fa63 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.756111] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Uploading image cb093f7a-5b17-4099-b4b6-8313549073b5 {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 976.821735] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 976.821735] env[62814]: value = "vm-845767" [ 976.821735] env[62814]: _type = "VirtualMachine" [ 976.821735] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 976.822040] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-b6bd7195-a54f-4ad4-b8f6-1599ab56c98c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.833387] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lease: (returnval){ [ 976.833387] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd1c48-9d4b-f686-620c-9e712f2b94c4" [ 976.833387] env[62814]: _type = "HttpNfcLease" [ 976.833387] env[62814]: } obtained for exporting VM: (result){ [ 976.833387] env[62814]: value = "vm-845767" [ 976.833387] env[62814]: _type = "VirtualMachine" [ 976.833387] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 976.837293] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the lease: (returnval){ [ 976.837293] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd1c48-9d4b-f686-620c-9e712f2b94c4" [ 976.837293] env[62814]: _type = "HttpNfcLease" [ 976.837293] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 976.849285] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 976.849285] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd1c48-9d4b-f686-620c-9e712f2b94c4" [ 976.849285] env[62814]: _type = "HttpNfcLease" [ 976.849285] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 976.879964] env[62814]: DEBUG oslo_vmware.api [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294135, 'name': PowerOnVM_Task, 'duration_secs': 0.511518} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.880319] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 976.880536] env[62814]: INFO nova.compute.manager [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Took 8.79 seconds to spawn the instance on the hypervisor. [ 976.880718] env[62814]: DEBUG nova.compute.manager [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 976.881626] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e19bf33-7769-44f8-bde6-0073c9397f92 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.962311] env[62814]: DEBUG oslo_concurrency.lockutils [None req-aa144261-0527-4b0c-81d6-dd8b40c9bb7d tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "40e2d845-0211-4c84-aef7-94014f999e1d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.428s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.177932] env[62814]: DEBUG oslo_concurrency.lockutils [None req-22f1c4af-be59-4961-8332-fa6c52d31aea tempest-VolumesAdminNegativeTest-1202358734 tempest-VolumesAdminNegativeTest-1202358734-project-member] Lock "5c574786-64a4-4f07-a267-101ecaaa6938" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.647s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 977.348024] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 977.348024] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd1c48-9d4b-f686-620c-9e712f2b94c4" [ 977.348024] env[62814]: _type = "HttpNfcLease" [ 977.348024] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 977.348024] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 977.348024] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd1c48-9d4b-f686-620c-9e712f2b94c4" [ 977.348024] env[62814]: _type = "HttpNfcLease" [ 977.348024] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 977.348024] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e80819-8e59-4ccd-a408-9a44c14ae58c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.359223] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5989d-96dd-eb94-8745-ec32d1591ebf/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 977.359223] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5989d-96dd-eb94-8745-ec32d1591ebf/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 977.428255] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26724df6-0ef6-4c49-8001-42db120ec1f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.436433] env[62814]: INFO nova.compute.manager [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Took 36.86 seconds to build instance. [ 977.440858] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83d983c-20fb-414c-aa88-3a156d278e6b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.479395] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddba7b28-5319-4769-a0a7-8815211f74ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.492296] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f004071e-6f64-4126-afbc-91ed7d4615ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.497801] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-116c352c-ae72-454c-8f20-6ca093ef06b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.514527] env[62814]: DEBUG nova.compute.provider_tree [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 977.657784] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 977.657784] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b17ad7ee-2f3a-451c-8280-6a136951b807 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.668164] env[62814]: DEBUG oslo_vmware.api [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 977.668164] env[62814]: value = "task-4294137" [ 977.668164] env[62814]: _type = "Task" [ 977.668164] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.681215] env[62814]: DEBUG oslo_vmware.api [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294137, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.940770] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6559206a-aedd-4cfb-8f47-295176e7cb22 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.378s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.190216] env[62814]: DEBUG oslo_vmware.api [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294137, 'name': PowerOffVM_Task, 'duration_secs': 0.230546} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.190216] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 978.190216] env[62814]: DEBUG nova.compute.manager [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 978.190216] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a19ffd5-6d90-4f4e-8537-ed38e9758b9c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.257700] env[62814]: DEBUG nova.scheduler.client.report [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 110 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 978.258105] env[62814]: DEBUG nova.compute.provider_tree [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 110 to 111 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 978.259022] env[62814]: DEBUG nova.compute.provider_tree [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 978.279206] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.279206] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.279206] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 978.279206] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.279206] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.280768] env[62814]: INFO nova.compute.manager [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Terminating instance [ 978.707766] env[62814]: DEBUG oslo_concurrency.lockutils [None req-13f81839-f0be-4ade-8f58-5db48f37de49 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.090s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.769617] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.116s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 978.771473] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.950s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 978.771601] env[62814]: DEBUG nova.objects.instance [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lazy-loading 'resources' on Instance uuid 844737d8-d852-44bb-bf9d-e673e737ef33 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.788235] env[62814]: DEBUG nova.compute.manager [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 978.788235] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.788615] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9097b883-dea9-44e6-ac55-6cfa777eb07b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.804307] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.804768] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ad40ff73-22ec-4cc0-b2fb-d4124efe9b07 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.813825] env[62814]: DEBUG oslo_vmware.api [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 978.813825] env[62814]: value = "task-4294138" [ 978.813825] env[62814]: _type = "Task" [ 978.813825] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.832594] env[62814]: DEBUG oslo_vmware.api [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294138, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.908158] env[62814]: INFO nova.scheduler.client.report [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Deleted allocations for instance 9be62576-5a05-473f-befd-b33f5fde9185 [ 979.327043] env[62814]: DEBUG oslo_vmware.api [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294138, 'name': PowerOffVM_Task, 'duration_secs': 0.275062} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.327417] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.327648] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 979.328402] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6bd6ba4d-c24a-40f7-b74c-2b86298c6378 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.402932] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 979.403195] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 979.403402] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Deleting the datastore file [datastore2] 3e4479a0-8edd-4b37-8cc9-2c91275b88ee {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.403700] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b85d7bd-96c0-4cd2-84f2-3761e6e2d6ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.418483] env[62814]: DEBUG oslo_vmware.api [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 979.418483] env[62814]: value = "task-4294140" [ 979.418483] env[62814]: _type = "Task" [ 979.418483] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.419066] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2af7e640-16d0-4066-a844-f7fb02969543 tempest-ServersWithSpecificFlavorTestJSON-2062133425 tempest-ServersWithSpecificFlavorTestJSON-2062133425-project-member] Lock "9be62576-5a05-473f-befd-b33f5fde9185" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.487s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 979.448382] env[62814]: DEBUG oslo_vmware.api [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.880512] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1650757-9bf5-4eed-b963-93c37349afc9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.894533] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a211d27f-3e82-4523-bd90-03f450ca3afd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.935156] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20508b5f-fb65-4941-8c81-3fe2f995d811 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.947087] env[62814]: DEBUG oslo_vmware.api [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294140, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.289382} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.947531] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.947786] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.948080] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.948319] env[62814]: INFO nova.compute.manager [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Took 1.16 seconds to destroy the instance on the hypervisor. [ 979.948641] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 979.950132] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5932d368-7b15-46e9-8330-8fd21f41bbb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.955293] env[62814]: DEBUG nova.compute.manager [-] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 979.955293] env[62814]: DEBUG nova.network.neutron [-] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 979.968250] env[62814]: DEBUG nova.compute.provider_tree [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 980.557308] env[62814]: DEBUG nova.scheduler.client.report [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 111 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 980.557308] env[62814]: DEBUG nova.compute.provider_tree [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 111 to 112 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 980.557308] env[62814]: DEBUG nova.compute.provider_tree [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 981.063703] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.292s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 981.066953] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.815s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 981.069365] env[62814]: INFO nova.compute.claims [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 981.211814] env[62814]: INFO nova.scheduler.client.report [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Deleted allocations for instance 844737d8-d852-44bb-bf9d-e673e737ef33 [ 981.725282] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c88bd73-6fc3-4b3f-8ca3-46537ed9b586 tempest-ServersTestManualDisk-1312769716 tempest-ServersTestManualDisk-1312769716-project-member] Lock "844737d8-d852-44bb-bf9d-e673e737ef33" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.755s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 981.962493] env[62814]: DEBUG nova.network.neutron [-] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.465274] env[62814]: INFO nova.compute.manager [-] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Took 2.51 seconds to deallocate network for instance. [ 982.635802] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-586d096f-e060-4e8c-949d-863f31d42f66 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.644651] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a8a9a7-39d0-48c7-ae25-4337f3977785 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.684212] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3135489e-5aac-45d0-b352-7b4ab028cbce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.697091] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e0d83c-ba7a-4bee-b472-c8fdd2e341bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.706172] env[62814]: DEBUG nova.compute.provider_tree [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.863229] env[62814]: DEBUG oslo_concurrency.lockutils [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 982.863524] env[62814]: DEBUG oslo_concurrency.lockutils [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.863734] env[62814]: DEBUG oslo_concurrency.lockutils [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 982.863917] env[62814]: DEBUG oslo_concurrency.lockutils [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 982.864091] env[62814]: DEBUG oslo_concurrency.lockutils [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 982.868947] env[62814]: INFO nova.compute.manager [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Terminating instance [ 982.981954] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 983.214375] env[62814]: DEBUG nova.scheduler.client.report [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 983.373330] env[62814]: DEBUG nova.compute.manager [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 983.373757] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 983.374985] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4128506-a669-4769-824a-4796671cadee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.383792] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 983.384159] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ed5986b0-046d-4abc-8f02-cb8191d799e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.392392] env[62814]: DEBUG oslo_vmware.api [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 983.392392] env[62814]: value = "task-4294141" [ 983.392392] env[62814]: _type = "Task" [ 983.392392] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.401428] env[62814]: DEBUG oslo_vmware.api [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294141, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.719533] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.652s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 983.720222] env[62814]: DEBUG nova.compute.manager [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 983.722896] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.689s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 983.725420] env[62814]: INFO nova.compute.claims [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.904193] env[62814]: DEBUG oslo_vmware.api [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294141, 'name': PowerOffVM_Task, 'duration_secs': 0.327631} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.904576] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.904664] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 983.904931] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0f5a269-c770-4f0e-ad1e-208eb5c68bce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.975929] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 983.975929] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 983.975929] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleting the datastore file [datastore2] 98ba2786-023f-4d36-bdd0-e38bdd23d73c {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.976183] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c04ea2b-f544-4a45-868a-2efe0691d122 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.984162] env[62814]: DEBUG oslo_vmware.api [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 983.984162] env[62814]: value = "task-4294143" [ 983.984162] env[62814]: _type = "Task" [ 983.984162] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.993463] env[62814]: DEBUG oslo_vmware.api [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294143, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.229042] env[62814]: DEBUG nova.compute.utils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 984.233627] env[62814]: DEBUG nova.compute.manager [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 984.233825] env[62814]: DEBUG nova.network.neutron [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 984.320748] env[62814]: DEBUG nova.policy [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '58359539fd4545cdbf200ec364e43834', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3dedbeff832a4ac48b0aa01bd6acc3f1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 984.495433] env[62814]: DEBUG oslo_vmware.api [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294143, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.366112} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.495744] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 984.496141] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 984.496358] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 984.496596] env[62814]: INFO nova.compute.manager [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Took 1.12 seconds to destroy the instance on the hypervisor. [ 984.496851] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 984.497064] env[62814]: DEBUG nova.compute.manager [-] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 984.497164] env[62814]: DEBUG nova.network.neutron [-] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 984.738227] env[62814]: DEBUG nova.compute.manager [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 985.199879] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76767f6-eb12-46f5-a074-c7ddf906b3ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.216637] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb71d4a0-84b1-4d76-9c07-6c61ec56deba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.269061] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883ec8d8-4537-43f6-b48b-9404f0f34056 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.269569] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34dbbb1c-e5b5-4710-9e13-895effacf186 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.284414] env[62814]: DEBUG nova.compute.provider_tree [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.286571] env[62814]: DEBUG nova.network.neutron [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Successfully created port: 26bd8c43-7afd-47eb-b53a-52d4a7847cec {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 985.728480] env[62814]: DEBUG oslo_vmware.rw_handles [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5282c603-1bc8-2fe0-87ac-d4abf2f60ef2/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 985.729909] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcef9ca-bc52-46c5-ba07-35fdf1717b52 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.740882] env[62814]: DEBUG oslo_vmware.rw_handles [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5282c603-1bc8-2fe0-87ac-d4abf2f60ef2/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 985.740882] env[62814]: ERROR oslo_vmware.rw_handles [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5282c603-1bc8-2fe0-87ac-d4abf2f60ef2/disk-0.vmdk due to incomplete transfer. [ 985.740882] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-404d3711-d53e-4d62-bc77-00ac8a40a906 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.748430] env[62814]: DEBUG oslo_vmware.rw_handles [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5282c603-1bc8-2fe0-87ac-d4abf2f60ef2/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 985.748773] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Uploaded image e7d28230-ebf8-44c9-b3b7-03c29d4f076a to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 985.751524] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 985.751817] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-29ea2442-2db5-4716-9a72-47b9b366945f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.758077] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 985.758077] env[62814]: value = "task-4294144" [ 985.758077] env[62814]: _type = "Task" [ 985.758077] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.764922] env[62814]: DEBUG nova.compute.manager [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 985.774577] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294144, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.792561] env[62814]: DEBUG nova.scheduler.client.report [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 985.845405] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 985.845690] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.845859] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 985.846057] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.846245] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 985.846377] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 985.846606] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 985.847172] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 985.847172] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 985.847172] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 985.847319] env[62814]: DEBUG nova.virt.hardware [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 985.848259] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee878ec-5ce8-4a41-9402-11c920ed7c53 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.856537] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e9613a-1ce9-4184-869a-55d011adc804 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.038191] env[62814]: DEBUG nova.compute.manager [req-c9478022-1718-4512-a005-3629dd9b4ddb req-29fab917-da4c-4be8-8759-70e63878fa87 service nova] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Received event network-vif-deleted-95b486fb-bbe0-4666-a926-01bcfa37eaad {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 986.271618] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294144, 'name': Destroy_Task, 'duration_secs': 0.356064} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.272048] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Destroyed the VM [ 986.272412] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 986.272732] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cf37f116-38c0-431e-93f7-7b2fb09281c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.281455] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 986.281455] env[62814]: value = "task-4294145" [ 986.281455] env[62814]: _type = "Task" [ 986.281455] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.292708] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294145, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.296181] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.573s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 986.296824] env[62814]: DEBUG nova.compute.manager [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 986.301895] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.097s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 986.302314] env[62814]: DEBUG nova.objects.instance [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lazy-loading 'resources' on Instance uuid c8765756-0870-4a06-a1a5-d02177959b29 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.726939] env[62814]: DEBUG nova.objects.instance [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'flavor' on Instance uuid 9683894b-a300-4400-a1b9-db62478f42c5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.793607] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294145, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.807952] env[62814]: DEBUG nova.compute.utils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 986.812299] env[62814]: DEBUG nova.compute.manager [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 986.812299] env[62814]: DEBUG nova.network.neutron [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 986.821982] env[62814]: DEBUG nova.network.neutron [-] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.837345] env[62814]: DEBUG nova.compute.manager [req-38f05d9b-1beb-405b-9f50-b9579dfdffbe req-1d3b69db-696f-4e9f-a366-0dddcbe951df service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Received event network-changed-029acc9a-5528-4c96-a5a7-fa02e89a2bb9 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 986.837532] env[62814]: DEBUG nova.compute.manager [req-38f05d9b-1beb-405b-9f50-b9579dfdffbe req-1d3b69db-696f-4e9f-a366-0dddcbe951df service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Refreshing instance network info cache due to event network-changed-029acc9a-5528-4c96-a5a7-fa02e89a2bb9. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 986.837771] env[62814]: DEBUG oslo_concurrency.lockutils [req-38f05d9b-1beb-405b-9f50-b9579dfdffbe req-1d3b69db-696f-4e9f-a366-0dddcbe951df service nova] Acquiring lock "refresh_cache-40e2d845-0211-4c84-aef7-94014f999e1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.837913] env[62814]: DEBUG oslo_concurrency.lockutils [req-38f05d9b-1beb-405b-9f50-b9579dfdffbe req-1d3b69db-696f-4e9f-a366-0dddcbe951df service nova] Acquired lock "refresh_cache-40e2d845-0211-4c84-aef7-94014f999e1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 986.838685] env[62814]: DEBUG nova.network.neutron [req-38f05d9b-1beb-405b-9f50-b9579dfdffbe req-1d3b69db-696f-4e9f-a366-0dddcbe951df service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Refreshing network info cache for port 029acc9a-5528-4c96-a5a7-fa02e89a2bb9 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 986.878118] env[62814]: DEBUG nova.policy [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5779eaa60ec44b2b80660e521b598eb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10018060297c409d82935ad626cc54c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 987.235545] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.235545] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquired lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 987.235545] env[62814]: DEBUG nova.network.neutron [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 987.235545] env[62814]: DEBUG nova.objects.instance [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'info_cache' on Instance uuid 9683894b-a300-4400-a1b9-db62478f42c5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.293294] env[62814]: DEBUG oslo_vmware.api [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294145, 'name': RemoveSnapshot_Task, 'duration_secs': 0.807683} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.293658] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 987.293963] env[62814]: INFO nova.compute.manager [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Took 18.53 seconds to snapshot the instance on the hypervisor. [ 987.315243] env[62814]: DEBUG nova.compute.manager [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 987.327520] env[62814]: INFO nova.compute.manager [-] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Took 2.83 seconds to deallocate network for instance. [ 987.678116] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea5b5f1-00ca-4214-8ada-c456d8c45206 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.678116] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2034f0c6-aa21-4d0a-a59b-80e59df757b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.678116] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5989d-96dd-eb94-8745-ec32d1591ebf/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 987.678116] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56a9beea-4ec3-4eb2-9099-9eadffb471ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.678116] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8580e3b7-1291-471a-826e-17cd0fb19dc9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.683273] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5989d-96dd-eb94-8745-ec32d1591ebf/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 987.683273] env[62814]: ERROR oslo_vmware.rw_handles [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5989d-96dd-eb94-8745-ec32d1591ebf/disk-0.vmdk due to incomplete transfer. [ 987.683273] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1b172dc7-09c9-4ed6-99ca-3e6eb2b2d9b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.688985] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78710e19-754c-4943-9799-4196f767a36c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.694759] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d5989d-96dd-eb94-8745-ec32d1591ebf/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 987.694838] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Uploaded image cb093f7a-5b17-4099-b4b6-8313549073b5 to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 987.696503] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 987.697475] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8c6b6d89-95bf-4f4c-b7e9-a4da7ab55203 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.707291] env[62814]: DEBUG nova.compute.provider_tree [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.713285] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 987.713285] env[62814]: value = "task-4294146" [ 987.713285] env[62814]: _type = "Task" [ 987.713285] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.723141] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294146, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.737088] env[62814]: DEBUG nova.objects.base [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Object Instance<9683894b-a300-4400-a1b9-db62478f42c5> lazy-loaded attributes: flavor,info_cache {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 987.837658] env[62814]: DEBUG oslo_concurrency.lockutils [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 987.882452] env[62814]: DEBUG nova.compute.manager [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Found 3 images (rotation: 2) {{(pid=62814) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5017}} [ 987.882662] env[62814]: DEBUG nova.compute.manager [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Rotating out 1 backups {{(pid=62814) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5025}} [ 987.882827] env[62814]: DEBUG nova.compute.manager [None req-74d57ed9-4525-4351-ba85-f68d6cef8299 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Deleting image af88e617-86b6-4046-9983-8f685578152f {{(pid=62814) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5030}} [ 988.213965] env[62814]: DEBUG nova.scheduler.client.report [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 988.225685] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294146, 'name': Destroy_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.409306] env[62814]: INFO nova.compute.manager [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Rebuilding instance [ 988.669512] env[62814]: DEBUG nova.compute.manager [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 988.672249] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f45d141-df7e-4c9f-8c5a-5fda458864de {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.729036] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.427s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 988.731100] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294146, 'name': Destroy_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.731774] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.928s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 988.731774] env[62814]: DEBUG nova.objects.instance [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lazy-loading 'resources' on Instance uuid f5ad9d70-75fb-4881-8853-5ede4d0903f2 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.794703] env[62814]: DEBUG nova.network.neutron [req-38f05d9b-1beb-405b-9f50-b9579dfdffbe req-1d3b69db-696f-4e9f-a366-0dddcbe951df service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Updated VIF entry in instance network info cache for port 029acc9a-5528-4c96-a5a7-fa02e89a2bb9. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 988.795118] env[62814]: DEBUG nova.network.neutron [req-38f05d9b-1beb-405b-9f50-b9579dfdffbe req-1d3b69db-696f-4e9f-a366-0dddcbe951df service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Updating instance_info_cache with network_info: [{"id": "029acc9a-5528-4c96-a5a7-fa02e89a2bb9", "address": "fa:16:3e:e8:ed:4e", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.215", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap029acc9a-55", "ovs_interfaceid": "029acc9a-5528-4c96-a5a7-fa02e89a2bb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.828852] env[62814]: DEBUG nova.compute.manager [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 988.846097] env[62814]: INFO nova.scheduler.client.report [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted allocations for instance c8765756-0870-4a06-a1a5-d02177959b29 [ 988.917440] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 988.918641] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 988.918955] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 988.919396] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 988.919657] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 988.922025] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 988.922025] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 988.922025] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 988.922025] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 988.922025] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 988.922025] env[62814]: DEBUG nova.virt.hardware [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 988.922025] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1f8127-84b9-4807-96f0-e794f5bc6dc9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.938782] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fa885b-6d6a-4f2b-941e-8852e874e03a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.175035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 989.175035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 989.234115] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294146, 'name': Destroy_Task, 'duration_secs': 1.333223} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.234115] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Destroyed the VM [ 989.234115] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 989.234247] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4cc61c24-43c8-4fcf-94ce-801f82bf37d6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.243016] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 989.243016] env[62814]: value = "task-4294147" [ 989.243016] env[62814]: _type = "Task" [ 989.243016] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.259558] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294147, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.298176] env[62814]: DEBUG oslo_concurrency.lockutils [req-38f05d9b-1beb-405b-9f50-b9579dfdffbe req-1d3b69db-696f-4e9f-a366-0dddcbe951df service nova] Releasing lock "refresh_cache-40e2d845-0211-4c84-aef7-94014f999e1d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 989.331896] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 989.332201] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 989.684243] env[62814]: DEBUG nova.compute.manager [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 989.691010] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 989.691570] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7b1cee6-4e44-4ef7-aadc-348d1e080a24 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.700269] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 989.700269] env[62814]: value = "task-4294148" [ 989.700269] env[62814]: _type = "Task" [ 989.700269] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.708192] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.760110] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294147, 'name': RemoveSnapshot_Task} progress is 44%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.760110] env[62814]: DEBUG nova.network.neutron [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Successfully created port: 75b18381-f5ef-4ce2-ab71-634c787eaffe {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.852801] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 989.852801] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 989.852801] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 989.852801] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 989.852801] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 989.852801] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 989.853478] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 989.853478] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 989.869590] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8fd4eb6d-3cf0-4442-83f8-916f88311d6f tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "c8765756-0870-4a06-a1a5-d02177959b29" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.238s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 990.095021] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1f2af6-d95c-4368-b326-87426087c2e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.108797] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e47655-9cf6-4d46-bb24-a58fa159dd1f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.146595] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41eadb4-0455-42cd-8aac-aca28e7b3d29 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.156322] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77dc1ca6-314e-41d8-9fa1-3118a07dc202 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.182717] env[62814]: DEBUG nova.compute.provider_tree [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.195962] env[62814]: DEBUG nova.compute.manager [req-37767043-fbfd-4b20-941c-651619388143 req-e2ba4242-831b-4114-89f8-109db6891989 service nova] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Received event network-vif-deleted-742fe6f0-d787-4ca8-be9a-edabfb0646a0 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 990.222493] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294148, 'name': PowerOffVM_Task, 'duration_secs': 0.186783} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.222493] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 990.222493] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 990.222493] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec756a94-a253-4120-b3c6-d41cc3d0dec6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.228826] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 990.229058] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-10902451-1a60-4947-97df-aa508b1edad2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.253538] env[62814]: DEBUG oslo_vmware.api [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294147, 'name': RemoveSnapshot_Task, 'duration_secs': 0.836798} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.253787] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 990.254014] env[62814]: INFO nova.compute.manager [None req-7a4be570-4d86-4faa-bbac-15a107907dba tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Took 18.76 seconds to snapshot the instance on the hypervisor. [ 990.280981] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 990.293475] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 990.293692] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 990.293902] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleting the datastore file [datastore2] b5ddd05f-2027-4edf-84c5-0d2f537a95e4 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.294211] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7b0ee5ba-0536-461e-a7c8-558e74cbad17 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.302043] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 990.302043] env[62814]: value = "task-4294150" [ 990.302043] env[62814]: _type = "Task" [ 990.302043] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.316740] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294150, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.360261] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 990.443258] env[62814]: DEBUG nova.network.neutron [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Updating instance_info_cache with network_info: [{"id": "9fed46ec-04e6-4482-a460-370e4ff04a32", "address": "fa:16:3e:19:a2:ac", "network": {"id": "5363651c-6ade-4a4c-b762-9fc7cbbcf5ea", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1833251571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b3ba1c9c7624abfb3e0bd880dc40166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fed46ec-04", "ovs_interfaceid": "9fed46ec-04e6-4482-a460-370e4ff04a32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.465866] env[62814]: DEBUG nova.network.neutron [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Successfully updated port: 26bd8c43-7afd-47eb-b53a-52d4a7847cec {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 990.688568] env[62814]: DEBUG nova.scheduler.client.report [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 990.817702] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294150, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148108} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.817702] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 990.817702] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 990.817702] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 990.954496] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Releasing lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 990.972407] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "refresh_cache-fd1867e3-2523-4969-a157-b14c650f3779" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.972407] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "refresh_cache-fd1867e3-2523-4969-a157-b14c650f3779" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 990.972407] env[62814]: DEBUG nova.network.neutron [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 991.194427] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.463s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 991.197692] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.970s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 991.197692] env[62814]: DEBUG nova.objects.instance [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lazy-loading 'resources' on Instance uuid 22429810-3ea7-4472-9a90-018d6efcba1d {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.375528] env[62814]: INFO nova.scheduler.client.report [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleted allocations for instance f5ad9d70-75fb-4881-8853-5ede4d0903f2 [ 991.597486] env[62814]: DEBUG nova.network.neutron [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 991.891456] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 991.891832] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.891872] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 991.892084] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.892229] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 991.892404] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 991.892619] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 991.892777] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 991.893099] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 991.894032] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 991.894032] env[62814]: DEBUG nova.virt.hardware [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 991.894455] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17d45af-a230-469b-9a7b-6e145773d4a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.900420] env[62814]: DEBUG nova.network.neutron [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Updating instance_info_cache with network_info: [{"id": "26bd8c43-7afd-47eb-b53a-52d4a7847cec", "address": "fa:16:3e:a2:d7:1a", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26bd8c43-7a", "ovs_interfaceid": "26bd8c43-7afd-47eb-b53a-52d4a7847cec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.913497] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0e5aa0-320c-4a4d-bd14-ba26d7c3da59 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.926138] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:be:4b:c0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '605f83bd-808c-4b54-922e-54b14690987a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '260ddd77-4e00-4b1e-9299-fa71a7d23a57', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 991.933659] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 991.939404] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 991.939404] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b3ca617-ddf0-45b9-926b-272b3c449f36 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.958673] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 991.958673] env[62814]: value = "task-4294151" [ 991.958673] env[62814]: _type = "Task" [ 991.958673] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.959243] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 991.962606] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-550a0f68-51d7-47fe-8867-46eee33b11cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.974372] env[62814]: DEBUG oslo_vmware.api [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 991.974372] env[62814]: value = "task-4294152" [ 991.974372] env[62814]: _type = "Task" [ 991.974372] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.974372] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294151, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.984365] env[62814]: DEBUG oslo_vmware.api [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.323686] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67226c6c-970d-4d0d-9224-fabb0a793945 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.331783] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e0b711-3f8e-42d2-9086-27bebea6d093 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.362122] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67259ede-1cb2-460d-ac54-c6e8397dd7fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.369818] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb7d27b-d08c-4354-b3bc-92baae6c01bb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.387865] env[62814]: DEBUG nova.compute.provider_tree [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.389363] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7a64dc21-7a82-4a20-9be0-5eb71a8b116c tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "f5ad9d70-75fb-4881-8853-5ede4d0903f2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.141s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 992.405202] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "refresh_cache-fd1867e3-2523-4969-a157-b14c650f3779" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 992.405401] env[62814]: DEBUG nova.compute.manager [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Instance network_info: |[{"id": "26bd8c43-7afd-47eb-b53a-52d4a7847cec", "address": "fa:16:3e:a2:d7:1a", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26bd8c43-7a", "ovs_interfaceid": "26bd8c43-7afd-47eb-b53a-52d4a7847cec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 992.405883] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:d7:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3db2ab9e-1244-4377-b05f-ab76003f2428', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '26bd8c43-7afd-47eb-b53a-52d4a7847cec', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 992.413821] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 992.414330] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 992.414597] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b2186dd0-b36e-42ad-8699-b97bebcdada7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.434747] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 992.434747] env[62814]: value = "task-4294153" [ 992.434747] env[62814]: _type = "Task" [ 992.434747] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.442597] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294153, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.467941] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294151, 'name': CreateVM_Task, 'duration_secs': 0.361758} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.468142] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 992.468811] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.468997] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 992.469406] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 992.469664] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a008e392-466c-4331-9b8b-01872d1011f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.474032] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 992.474032] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52852826-8c86-932a-bc6c-c86015cbcd18" [ 992.474032] env[62814]: _type = "Task" [ 992.474032] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.487379] env[62814]: DEBUG oslo_vmware.api [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294152, 'name': PowerOnVM_Task, 'duration_secs': 0.458665} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.490068] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 992.490318] env[62814]: DEBUG nova.compute.manager [None req-ba0444b6-e941-4f7d-9068-717b2c2f4134 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 992.490635] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52852826-8c86-932a-bc6c-c86015cbcd18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.501489] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff292bd-c0f3-457c-ae9b-963a7b1542cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.893758] env[62814]: DEBUG nova.scheduler.client.report [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 992.949688] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294153, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.995811] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52852826-8c86-932a-bc6c-c86015cbcd18, 'name': SearchDatastore_Task, 'duration_secs': 0.012089} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.996232] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 992.996500] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 992.996756] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.996903] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 992.997095] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 992.997383] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-538a34f0-6dbd-45cd-8155-c8e144925806 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.008437] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 993.008657] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 993.013035] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4134934b-4a75-48c8-9275-a7174c8eebb7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.023045] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 993.023045] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5288dc34-8c7f-3ccc-50bc-f53193dfed78" [ 993.023045] env[62814]: _type = "Task" [ 993.023045] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.033757] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5288dc34-8c7f-3ccc-50bc-f53193dfed78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.313569] env[62814]: DEBUG nova.network.neutron [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Successfully updated port: 75b18381-f5ef-4ce2-ab71-634c787eaffe {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.400115] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.203s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 993.403063] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.175s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 993.403307] env[62814]: DEBUG nova.objects.instance [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lazy-loading 'resources' on Instance uuid 267c1ef2-ccc3-470e-943d-5fe9203c4dbc {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 993.446195] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294153, 'name': CreateVM_Task, 'duration_secs': 0.64215} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.447110] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 993.447855] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.448034] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 993.448377] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 993.450178] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d14e933-6203-40de-94df-3016ca107e58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.454969] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 993.454969] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5232df9c-292f-1d2d-c0a5-70f5fdd19fd9" [ 993.454969] env[62814]: _type = "Task" [ 993.454969] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.465842] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5232df9c-292f-1d2d-c0a5-70f5fdd19fd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.509153] env[62814]: INFO nova.scheduler.client.report [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted allocations for instance 22429810-3ea7-4472-9a90-018d6efcba1d [ 993.534564] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5288dc34-8c7f-3ccc-50bc-f53193dfed78, 'name': SearchDatastore_Task, 'duration_secs': 0.016908} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.535391] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94dcfa19-79be-41c6-937e-6302ed2cc1a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.540966] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 993.540966] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524efa19-88e7-6192-c39a-50b6b2855304" [ 993.540966] env[62814]: _type = "Task" [ 993.540966] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.551577] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524efa19-88e7-6192-c39a-50b6b2855304, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.559041] env[62814]: DEBUG nova.compute.manager [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Received event network-vif-plugged-26bd8c43-7afd-47eb-b53a-52d4a7847cec {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 993.559041] env[62814]: DEBUG oslo_concurrency.lockutils [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] Acquiring lock "fd1867e3-2523-4969-a157-b14c650f3779-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 993.559041] env[62814]: DEBUG oslo_concurrency.lockutils [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] Lock "fd1867e3-2523-4969-a157-b14c650f3779-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.003s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 993.559041] env[62814]: DEBUG oslo_concurrency.lockutils [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] Lock "fd1867e3-2523-4969-a157-b14c650f3779-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 993.559041] env[62814]: DEBUG nova.compute.manager [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] [instance: fd1867e3-2523-4969-a157-b14c650f3779] No waiting events found dispatching network-vif-plugged-26bd8c43-7afd-47eb-b53a-52d4a7847cec {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 993.559041] env[62814]: WARNING nova.compute.manager [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Received unexpected event network-vif-plugged-26bd8c43-7afd-47eb-b53a-52d4a7847cec for instance with vm_state building and task_state spawning. [ 993.559041] env[62814]: DEBUG nova.compute.manager [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Received event network-changed-26bd8c43-7afd-47eb-b53a-52d4a7847cec {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 993.559041] env[62814]: DEBUG nova.compute.manager [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Refreshing instance network info cache due to event network-changed-26bd8c43-7afd-47eb-b53a-52d4a7847cec. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 993.559041] env[62814]: DEBUG oslo_concurrency.lockutils [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] Acquiring lock "refresh_cache-fd1867e3-2523-4969-a157-b14c650f3779" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.559041] env[62814]: DEBUG oslo_concurrency.lockutils [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] Acquired lock "refresh_cache-fd1867e3-2523-4969-a157-b14c650f3779" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 993.559041] env[62814]: DEBUG nova.network.neutron [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Refreshing network info cache for port 26bd8c43-7afd-47eb-b53a-52d4a7847cec {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 993.817954] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.817954] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 993.818240] env[62814]: DEBUG nova.network.neutron [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 993.968318] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5232df9c-292f-1d2d-c0a5-70f5fdd19fd9, 'name': SearchDatastore_Task, 'duration_secs': 0.022423} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.968658] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 993.968880] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 993.969147] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.026247] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6afb1458-03f9-4737-905e-e4d4a89d1072 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "22429810-3ea7-4472-9a90-018d6efcba1d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.195s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 994.054896] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524efa19-88e7-6192-c39a-50b6b2855304, 'name': SearchDatastore_Task, 'duration_secs': 0.027879} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.055185] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 994.055494] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b5ddd05f-2027-4edf-84c5-0d2f537a95e4/b5ddd05f-2027-4edf-84c5-0d2f537a95e4.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 994.058048] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 994.058659] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.058659] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1474ac37-c4ae-493b-92c2-a4d727125348 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.060485] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dbe9796-1290-4473-902b-f04de6bd394a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.072694] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.072865] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 994.074586] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aefc3ffa-47dd-4d80-9252-a1e1d2afb62d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.077157] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 994.077157] env[62814]: value = "task-4294154" [ 994.077157] env[62814]: _type = "Task" [ 994.077157] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.087064] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 994.087064] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52309808-c0ea-12e3-3e60-360fcc1f40cc" [ 994.087064] env[62814]: _type = "Task" [ 994.087064] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.096943] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294154, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.096943] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52309808-c0ea-12e3-3e60-360fcc1f40cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.401945] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18639515-8c60-45c5-a7bb-387c6eac875f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.409616] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db674a13-e85f-42b8-849c-b8a55d5b81c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.444637] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-995f8180-06a0-46d9-8a22-fd5e2d678f81 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.454103] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f56be7a-0e86-4ffe-a27a-9816d856bf18 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.469802] env[62814]: DEBUG nova.compute.provider_tree [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.471865] env[62814]: DEBUG nova.network.neutron [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 994.596408] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294154, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.602334] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52309808-c0ea-12e3-3e60-360fcc1f40cc, 'name': SearchDatastore_Task, 'duration_secs': 0.028965} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.603170] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78618581-73df-4ba1-bc0f-f8441ac93245 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.607920] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 994.607920] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52456371-8c88-602f-4fd3-298d83a0c1b8" [ 994.607920] env[62814]: _type = "Task" [ 994.607920] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.615616] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52456371-8c88-602f-4fd3-298d83a0c1b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.826182] env[62814]: DEBUG nova.network.neutron [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Updated VIF entry in instance network info cache for port 26bd8c43-7afd-47eb-b53a-52d4a7847cec. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 994.826472] env[62814]: DEBUG nova.network.neutron [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Updating instance_info_cache with network_info: [{"id": "26bd8c43-7afd-47eb-b53a-52d4a7847cec", "address": "fa:16:3e:a2:d7:1a", "network": {"id": "37e61cfc-9573-498e-9e85-98dd8eb7e5a6", "bridge": "br-int", "label": "tempest-ImagesTestJSON-10724113-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.3"}}], "meta": {"injected": false, "tenant_id": "3dedbeff832a4ac48b0aa01bd6acc3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3db2ab9e-1244-4377-b05f-ab76003f2428", "external-id": "nsx-vlan-transportzone-199", "segmentation_id": 199, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26bd8c43-7a", "ovs_interfaceid": "26bd8c43-7afd-47eb-b53a-52d4a7847cec", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.831425] env[62814]: DEBUG nova.network.neutron [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updating instance_info_cache with network_info: [{"id": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "address": "fa:16:3e:ee:a8:e0", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b18381-f5", "ovs_interfaceid": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.975734] env[62814]: DEBUG nova.scheduler.client.report [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 995.092782] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294154, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.122794] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52456371-8c88-602f-4fd3-298d83a0c1b8, 'name': SearchDatastore_Task, 'duration_secs': 0.016758} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.122794] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 995.122794] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] fd1867e3-2523-4969-a157-b14c650f3779/fd1867e3-2523-4969-a157-b14c650f3779.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 995.122794] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7ce2d252-1757-4086-a8aa-fccf5fb97de7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.126737] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 995.126737] env[62814]: value = "task-4294155" [ 995.126737] env[62814]: _type = "Task" [ 995.126737] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.137892] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294155, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.331813] env[62814]: DEBUG oslo_concurrency.lockutils [req-1105fbf9-22f6-4f32-a15d-62f59bfabf54 req-fe78be6f-6d2d-4ef9-bc3f-2e7a7d51ed32 service nova] Releasing lock "refresh_cache-fd1867e3-2523-4969-a157-b14c650f3779" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 995.333581] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 995.333902] env[62814]: DEBUG nova.compute.manager [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Instance network_info: |[{"id": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "address": "fa:16:3e:ee:a8:e0", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b18381-f5", "ovs_interfaceid": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 995.334399] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:a8:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe99da4f-5630-4afd-918b-b327193d8489', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '75b18381-f5ef-4ce2-ab71-634c787eaffe', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.343971] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Creating folder: Project (10018060297c409d82935ad626cc54c9). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 995.344571] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-878e6030-f286-4acf-ab7f-d339966177bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.357516] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Created folder: Project (10018060297c409d82935ad626cc54c9) in parent group-v845547. [ 995.357727] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Creating folder: Instances. Parent ref: group-v845770. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 995.358036] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-718282e1-3e3d-4590-989e-46d3d5e7d0a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.367919] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Created folder: Instances in parent group-v845770. [ 995.367919] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 995.368121] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 995.368196] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e74ad2ae-e569-486d-be18-f712fdf46f3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.386843] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.386843] env[62814]: value = "task-4294158" [ 995.386843] env[62814]: _type = "Task" [ 995.386843] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.396150] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294158, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.484065] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.079s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.484385] env[62814]: DEBUG oslo_concurrency.lockutils [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.457s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.485145] env[62814]: DEBUG nova.objects.instance [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lazy-loading 'resources' on Instance uuid 6a592192-1b41-4be2-84a6-c3b76a4e5643 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.588019] env[62814]: INFO nova.scheduler.client.report [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted allocations for instance 267c1ef2-ccc3-470e-943d-5fe9203c4dbc [ 995.595371] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294154, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.174473} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.595895] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b5ddd05f-2027-4edf-84c5-0d2f537a95e4/b5ddd05f-2027-4edf-84c5-0d2f537a95e4.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 995.596132] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 995.596399] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf765731-af80-4529-937c-34ce126fcb3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.603257] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 995.603257] env[62814]: value = "task-4294159" [ 995.603257] env[62814]: _type = "Task" [ 995.603257] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.615846] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294159, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.640685] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294155, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.652109] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "4ed66b36-b6c6-4673-9c03-169a01134574" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.652300] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "4ed66b36-b6c6-4673-9c03-169a01134574" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.652527] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "4ed66b36-b6c6-4673-9c03-169a01134574-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 995.653042] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "4ed66b36-b6c6-4673-9c03-169a01134574-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 995.653042] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "4ed66b36-b6c6-4673-9c03-169a01134574-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 995.655668] env[62814]: INFO nova.compute.manager [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Terminating instance [ 995.897235] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294158, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.937030] env[62814]: DEBUG nova.compute.manager [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 995.937262] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8278748d-edff-40d1-896e-53e72134f66d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.101501] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b0ed6359-220c-4752-94ec-de5b7d40a473 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "267c1ef2-ccc3-470e-943d-5fe9203c4dbc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.821s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.113641] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294159, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.150308} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.113841] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.114758] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cab1ffe1-09fd-491a-b6ec-2cf5077852ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.140296] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] b5ddd05f-2027-4edf-84c5-0d2f537a95e4/b5ddd05f-2027-4edf-84c5-0d2f537a95e4.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.143818] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0b29b67-ea5c-49cd-b5cd-4098adde8feb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.162208] env[62814]: DEBUG nova.compute.manager [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 996.162413] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 996.164025] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ee41337-5642-42b9-88a0-0b208e419546 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.171305] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 996.171305] env[62814]: value = "task-4294160" [ 996.171305] env[62814]: _type = "Task" [ 996.171305] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.172317] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294155, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.910081} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.172445] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] fd1867e3-2523-4969-a157-b14c650f3779/fd1867e3-2523-4969-a157-b14c650f3779.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 996.172671] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.175496] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-66700764-f2de-4a6a-9e47-a00650a54b50 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.180546] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 996.180546] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc8938f7-f169-45fc-bbe2-ee9922a5c3d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.187555] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294160, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.187555] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 996.187555] env[62814]: value = "task-4294161" [ 996.187555] env[62814]: _type = "Task" [ 996.187555] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.190872] env[62814]: DEBUG oslo_vmware.api [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 996.190872] env[62814]: value = "task-4294162" [ 996.190872] env[62814]: _type = "Task" [ 996.190872] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.196846] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294161, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.201789] env[62814]: DEBUG oslo_vmware.api [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4294162, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.398360] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294158, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.448570] env[62814]: DEBUG nova.compute.manager [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Received event network-vif-plugged-75b18381-f5ef-4ce2-ab71-634c787eaffe {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 996.448570] env[62814]: DEBUG oslo_concurrency.lockutils [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] Acquiring lock "885e0798-2fa6-4f6b-82be-517b6d1168d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 996.448990] env[62814]: DEBUG oslo_concurrency.lockutils [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] Lock "885e0798-2fa6-4f6b-82be-517b6d1168d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 996.448990] env[62814]: DEBUG oslo_concurrency.lockutils [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] Lock "885e0798-2fa6-4f6b-82be-517b6d1168d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 996.450434] env[62814]: DEBUG nova.compute.manager [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] No waiting events found dispatching network-vif-plugged-75b18381-f5ef-4ce2-ab71-634c787eaffe {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 996.450434] env[62814]: WARNING nova.compute.manager [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Received unexpected event network-vif-plugged-75b18381-f5ef-4ce2-ab71-634c787eaffe for instance with vm_state building and task_state spawning. [ 996.450434] env[62814]: DEBUG nova.compute.manager [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Received event network-changed-75b18381-f5ef-4ce2-ab71-634c787eaffe {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 996.450434] env[62814]: DEBUG nova.compute.manager [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Refreshing instance network info cache due to event network-changed-75b18381-f5ef-4ce2-ab71-634c787eaffe. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 996.450434] env[62814]: DEBUG oslo_concurrency.lockutils [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] Acquiring lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.450434] env[62814]: DEBUG oslo_concurrency.lockutils [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] Acquired lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 996.450434] env[62814]: DEBUG nova.network.neutron [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Refreshing network info cache for port 75b18381-f5ef-4ce2-ab71-634c787eaffe {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 996.452670] env[62814]: INFO nova.compute.manager [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] instance snapshotting [ 996.455747] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a530c03f-3f82-425f-a511-389cc8dee188 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.482119] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936889d4-1cbd-4940-9400-67bceed97f5f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.526912] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e964602c-1580-4516-a764-1fd110f1f1ac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.533871] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a1e4df-5f8d-428d-adf1-ba011891e61c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.565865] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e1958f-d0c6-4a55-8bd8-b23162f48922 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.573828] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4fe1bb-e493-464e-9f17-655af5bc70ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.586981] env[62814]: DEBUG nova.compute.provider_tree [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 996.681388] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294160, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.698557] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294161, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.231508} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.700853] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.700853] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6720d741-10ff-46ab-8ef2-b5444c72c493 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.707128] env[62814]: DEBUG oslo_vmware.api [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4294162, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.725802] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] fd1867e3-2523-4969-a157-b14c650f3779/fd1867e3-2523-4969-a157-b14c650f3779.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.725802] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-416fcdc0-d4c1-46e8-b0a2-c50184dd2600 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.750885] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 996.750885] env[62814]: value = "task-4294163" [ 996.750885] env[62814]: _type = "Task" [ 996.750885] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.759623] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294163, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.898220] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294158, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.994080] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 996.994410] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-7eb37f86-efab-45b8-a811-077d6eb224f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.004150] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 997.004150] env[62814]: value = "task-4294164" [ 997.004150] env[62814]: _type = "Task" [ 997.004150] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.015408] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294164, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.091954] env[62814]: DEBUG nova.scheduler.client.report [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 997.184708] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294160, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.203331] env[62814]: DEBUG oslo_vmware.api [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4294162, 'name': PowerOffVM_Task, 'duration_secs': 0.643033} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.203575] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.204350] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 997.204350] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-110b185e-336e-4ea9-8fbe-a172c51fb422 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.260699] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294163, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.285350] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 997.285487] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 997.285621] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleting the datastore file [datastore2] 4ed66b36-b6c6-4673-9c03-169a01134574 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 997.285925] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f1f85556-80e5-4bc9-a5e8-ccaac5e9ba58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.294885] env[62814]: DEBUG oslo_vmware.api [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for the task: (returnval){ [ 997.294885] env[62814]: value = "task-4294166" [ 997.294885] env[62814]: _type = "Task" [ 997.294885] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.304120] env[62814]: DEBUG oslo_vmware.api [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4294166, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.402043] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294158, 'name': CreateVM_Task, 'duration_secs': 1.825553} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.402312] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 997.403222] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.403398] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 997.403728] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 997.403992] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d572f873-e729-43b5-bb0a-6ba633c599e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.410585] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 997.410585] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522ba0da-2dd7-b49f-c8f4-8190d61ea95a" [ 997.410585] env[62814]: _type = "Task" [ 997.410585] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.421896] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522ba0da-2dd7-b49f-c8f4-8190d61ea95a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.448595] env[62814]: DEBUG nova.network.neutron [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updated VIF entry in instance network info cache for port 75b18381-f5ef-4ce2-ab71-634c787eaffe. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 997.448595] env[62814]: DEBUG nova.network.neutron [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updating instance_info_cache with network_info: [{"id": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "address": "fa:16:3e:ee:a8:e0", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b18381-f5", "ovs_interfaceid": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.449763] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "9343301c-dfe9-41b0-b4a0-067af544d297" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 997.449990] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.450174] env[62814]: DEBUG nova.compute.manager [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 997.451174] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e28a99-a166-49a1-a902-bb80aef5d395 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.457606] env[62814]: DEBUG nova.compute.manager [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 997.458355] env[62814]: DEBUG nova.objects.instance [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'flavor' on Instance uuid 9343301c-dfe9-41b0-b4a0-067af544d297 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 997.512840] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294164, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.595820] env[62814]: DEBUG oslo_concurrency.lockutils [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.112s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 997.598380] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.097s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 997.599931] env[62814]: INFO nova.compute.claims [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 997.628973] env[62814]: INFO nova.scheduler.client.report [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Deleted allocations for instance 6a592192-1b41-4be2-84a6-c3b76a4e5643 [ 997.683453] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294160, 'name': ReconfigVM_Task, 'duration_secs': 1.234464} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.683882] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Reconfigured VM instance instance-0000004b to attach disk [datastore2] b5ddd05f-2027-4edf-84c5-0d2f537a95e4/b5ddd05f-2027-4edf-84c5-0d2f537a95e4.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.684747] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8332ad4f-f7dd-44b8-9b25-d9adc2ca241c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.690776] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 997.690776] env[62814]: value = "task-4294167" [ 997.690776] env[62814]: _type = "Task" [ 997.690776] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.699826] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294167, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.763225] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294163, 'name': ReconfigVM_Task, 'duration_secs': 0.916655} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.763225] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Reconfigured VM instance instance-0000004c to attach disk [datastore2] fd1867e3-2523-4969-a157-b14c650f3779/fd1867e3-2523-4969-a157-b14c650f3779.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.763832] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-01feabbd-3890-4165-8850-0cb0d4c28584 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.770407] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 997.770407] env[62814]: value = "task-4294168" [ 997.770407] env[62814]: _type = "Task" [ 997.770407] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.781532] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294168, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.812017] env[62814]: DEBUG oslo_vmware.api [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Task: {'id': task-4294166, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.475333} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.812392] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 997.812670] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 997.812917] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 997.813192] env[62814]: INFO nova.compute.manager [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Took 1.65 seconds to destroy the instance on the hypervisor. [ 997.813519] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 997.813765] env[62814]: DEBUG nova.compute.manager [-] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 997.813884] env[62814]: DEBUG nova.network.neutron [-] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 997.924657] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522ba0da-2dd7-b49f-c8f4-8190d61ea95a, 'name': SearchDatastore_Task, 'duration_secs': 0.038309} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.924958] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 997.925243] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 997.925572] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.925801] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 997.926042] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.926385] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-499d8572-1058-4c10-b3ba-ac9a4a6cc9ef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.937799] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 997.937979] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 997.938739] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37fb94a5-21c4-452a-887d-f163710127ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.944757] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 997.944757] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52181aae-436b-cab8-682a-b866a9bd3a09" [ 997.944757] env[62814]: _type = "Task" [ 997.944757] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.952897] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52181aae-436b-cab8-682a-b866a9bd3a09, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.954785] env[62814]: DEBUG oslo_concurrency.lockutils [req-dd6505d5-5cad-4e3d-9ae0-aae2e45f9a03 req-87159373-e107-4215-86f8-14468cced4ab service nova] Releasing lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 998.012466] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294164, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.137951] env[62814]: DEBUG oslo_concurrency.lockutils [None req-edb335e8-88c4-4d19-856c-634afd89a9ea tempest-MigrationsAdminTest-19785326 tempest-MigrationsAdminTest-19785326-project-member] Lock "6a592192-1b41-4be2-84a6-c3b76a4e5643" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.038s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 998.205754] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294167, 'name': Rename_Task, 'duration_secs': 0.445053} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.205754] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.205754] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3fabde1c-655c-4e66-8ea8-be5452c96516 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.209872] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 998.209872] env[62814]: value = "task-4294169" [ 998.209872] env[62814]: _type = "Task" [ 998.209872] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.221660] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294169, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.287029] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294168, 'name': Rename_Task, 'duration_secs': 0.447902} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.287384] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 998.287510] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ad69b4be-4ad3-458a-8b71-a872c6a3c8a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.295680] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 998.295680] env[62814]: value = "task-4294170" [ 998.295680] env[62814]: _type = "Task" [ 998.295680] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.307690] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294170, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.457876] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52181aae-436b-cab8-682a-b866a9bd3a09, 'name': SearchDatastore_Task, 'duration_secs': 0.015359} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.459515] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23ebd394-f6d6-4766-88b5-27019983dfe3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.465537] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 998.465950] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 998.465950] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52456479-5ac9-67f5-91d9-3c4a7967c4c2" [ 998.465950] env[62814]: _type = "Task" [ 998.465950] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.466134] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e9364f8-3a73-4e85-9568-e36a04945588 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.475467] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52456479-5ac9-67f5-91d9-3c4a7967c4c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.478890] env[62814]: DEBUG oslo_vmware.api [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 998.478890] env[62814]: value = "task-4294171" [ 998.478890] env[62814]: _type = "Task" [ 998.478890] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.484293] env[62814]: DEBUG oslo_vmware.api [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294171, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.520692] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294164, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.720434] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294169, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.806715] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294170, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.914810] env[62814]: DEBUG nova.network.neutron [-] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.980165] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52456479-5ac9-67f5-91d9-3c4a7967c4c2, 'name': SearchDatastore_Task, 'duration_secs': 0.037512} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.983675] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 998.984019] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 885e0798-2fa6-4f6b-82be-517b6d1168d2/885e0798-2fa6-4f6b-82be-517b6d1168d2.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 998.986922] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-052f3abb-477e-479a-9233-604131677d5a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.994442] env[62814]: DEBUG oslo_vmware.api [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294171, 'name': PowerOffVM_Task, 'duration_secs': 0.46664} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.995821] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.996123] env[62814]: DEBUG nova.compute.manager [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 998.996770] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 998.996770] env[62814]: value = "task-4294172" [ 998.996770] env[62814]: _type = "Task" [ 998.996770] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.997563] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68117889-e7e8-4443-b2e9-70f90b7429e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.014505] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294172, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.030506] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294164, 'name': CreateSnapshot_Task, 'duration_secs': 1.913836} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.030506] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 999.030506] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d21f1dd8-c31e-4e03-8289-683d223c3210 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.223592] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294169, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.275722] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593fa4f7-1ae4-47e6-b368-2690cb96c070 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.283424] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811e1452-6575-43f4-a65b-6103857c294a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.319853] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4d3fd2-816e-4533-99d4-bdb523d0da23 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.328040] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294170, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.331485] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089a44d4-5880-425d-a296-cc98e1a8c621 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.346082] env[62814]: DEBUG nova.compute.provider_tree [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 999.365400] env[62814]: DEBUG nova.compute.manager [req-e115ecf0-7050-4135-87fa-76d5bfceabf0 req-15ecccee-0e0f-4907-b1a1-47db5dee82d9 service nova] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Received event network-vif-deleted-2b7654ff-51f0-4a51-9b69-04d3352ddbee {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 999.410306] env[62814]: INFO nova.compute.manager [-] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Took 1.60 seconds to deallocate network for instance. [ 999.511382] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294172, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.526945] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6f4928c0-078e-4e90-a82c-77bf5d5268a0 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.077s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 999.552675] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 999.553059] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c8942a22-fa5d-4f51-a961-d427af1cf1c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.563531] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 999.563531] env[62814]: value = "task-4294173" [ 999.563531] env[62814]: _type = "Task" [ 999.563531] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.578745] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294173, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.726292] env[62814]: DEBUG oslo_vmware.api [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294169, 'name': PowerOnVM_Task, 'duration_secs': 1.268631} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.726499] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.727482] env[62814]: DEBUG nova.compute.manager [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 999.727956] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e7d247-bca0-4126-9319-ed9ac2324116 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.803091] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "316931ae-7a62-4bac-81e4-1fee9a36164e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.803437] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.837696] env[62814]: DEBUG oslo_vmware.api [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294170, 'name': PowerOnVM_Task, 'duration_secs': 1.267976} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.838375] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 999.838375] env[62814]: INFO nova.compute.manager [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Took 14.08 seconds to spawn the instance on the hypervisor. [ 999.838498] env[62814]: DEBUG nova.compute.manager [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 999.839733] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238d3237-3f00-424a-8206-31e871514541 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.858896] env[62814]: DEBUG nova.scheduler.client.report [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 999.871348] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "4ebc3280-64a7-4e9b-8385-2128357d3422" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 999.871691] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "4ebc3280-64a7-4e9b-8385-2128357d3422" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 999.917081] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.012918] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294172, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.078120] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294173, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.256591] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.306709] env[62814]: DEBUG nova.compute.manager [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1000.378699] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.778s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.378699] env[62814]: DEBUG nova.compute.manager [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1000.383036] env[62814]: DEBUG nova.compute.manager [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1000.386025] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.324s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1000.387380] env[62814]: INFO nova.compute.claims [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1000.394408] env[62814]: INFO nova.compute.manager [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Took 44.16 seconds to build instance. [ 1000.514360] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294172, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.204123} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.514807] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 885e0798-2fa6-4f6b-82be-517b6d1168d2/885e0798-2fa6-4f6b-82be-517b6d1168d2.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1000.515077] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1000.515678] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fff338ca-cab0-4acf-9a0f-9013f6934b83 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.526904] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1000.526904] env[62814]: value = "task-4294174" [ 1000.526904] env[62814]: _type = "Task" [ 1000.526904] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.536509] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294174, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.583089] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294173, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.840455] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1000.898680] env[62814]: DEBUG nova.compute.utils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1000.906268] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fdb9c368-c256-4f13-a5ca-0a85f8a709b2 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "fd1867e3-2523-4969-a157-b14c650f3779" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.686s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1000.913021] env[62814]: DEBUG nova.compute.manager [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1000.913021] env[62814]: DEBUG nova.network.neutron [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1000.940010] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1001.046232] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294174, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082765} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.046695] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1001.048022] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc611e70-2526-4ad2-be9c-78f822e806a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.096041] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 885e0798-2fa6-4f6b-82be-517b6d1168d2/885e0798-2fa6-4f6b-82be-517b6d1168d2.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1001.101378] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db1ddfab-a0de-49af-99bd-ed8a6ca40101 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.143621] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294173, 'name': CloneVM_Task} progress is 95%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.145688] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1001.145688] env[62814]: value = "task-4294175" [ 1001.145688] env[62814]: _type = "Task" [ 1001.145688] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.161655] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294175, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.333501] env[62814]: DEBUG nova.policy [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1973a37801a94dc9913eee0361c402fd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2c3ae0fcddbe4ee7913d590a06716163', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1001.408634] env[62814]: DEBUG nova.compute.manager [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1001.592921] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294173, 'name': CloneVM_Task, 'duration_secs': 2.026587} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.593071] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Created linked-clone VM from snapshot [ 1001.593938] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2295597a-1612-4a15-a569-e198a9331705 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.611269] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Uploading image 83b049c5-1828-43ae-9dc4-e43a6a030542 {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1001.654287] env[62814]: DEBUG oslo_vmware.rw_handles [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1001.654287] env[62814]: value = "vm-845774" [ 1001.654287] env[62814]: _type = "VirtualMachine" [ 1001.654287] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1001.658229] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-bf6fca2e-b97c-4eb7-bcb0-4e1d0ce37f75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.674151] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294175, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.681269] env[62814]: DEBUG oslo_vmware.rw_handles [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lease: (returnval){ [ 1001.681269] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524e18db-b596-5ed4-6fac-133dad0c9d65" [ 1001.681269] env[62814]: _type = "HttpNfcLease" [ 1001.681269] env[62814]: } obtained for exporting VM: (result){ [ 1001.681269] env[62814]: value = "vm-845774" [ 1001.681269] env[62814]: _type = "VirtualMachine" [ 1001.681269] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1001.681763] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the lease: (returnval){ [ 1001.681763] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524e18db-b596-5ed4-6fac-133dad0c9d65" [ 1001.681763] env[62814]: _type = "HttpNfcLease" [ 1001.681763] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1001.691755] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1001.691755] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524e18db-b596-5ed4-6fac-133dad0c9d65" [ 1001.691755] env[62814]: _type = "HttpNfcLease" [ 1001.691755] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1001.692199] env[62814]: DEBUG oslo_vmware.rw_handles [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1001.692199] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524e18db-b596-5ed4-6fac-133dad0c9d65" [ 1001.692199] env[62814]: _type = "HttpNfcLease" [ 1001.692199] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1001.693589] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09ff164-2637-402b-882e-6f6741021f5a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.706550] env[62814]: DEBUG oslo_vmware.rw_handles [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52453c89-9711-2b9b-df20-96f0485f5086/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1001.707034] env[62814]: DEBUG oslo_vmware.rw_handles [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52453c89-9711-2b9b-df20-96f0485f5086/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1001.849305] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d9732591-828b-4e91-9b01-780057e00d43 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.019514] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7282fa35-d680-4a81-8649-7f3d15554090 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.029042] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-711983c1-7c81-44e2-9623-6461d9f85899 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.074632] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c289a126-1e71-4dd2-af05-303cc1cead3f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.082808] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a178412c-1e47-4318-8fbd-ace75a6ae5b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.102250] env[62814]: DEBUG nova.compute.provider_tree [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.164364] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294175, 'name': ReconfigVM_Task, 'duration_secs': 0.72059} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.164364] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 885e0798-2fa6-4f6b-82be-517b6d1168d2/885e0798-2fa6-4f6b-82be-517b6d1168d2.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1002.165711] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49d964d4-9cb2-4e33-b8b5-a308a9166879 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.174041] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1002.174041] env[62814]: value = "task-4294177" [ 1002.174041] env[62814]: _type = "Task" [ 1002.174041] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.183944] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294177, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.418033] env[62814]: DEBUG nova.compute.manager [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Stashing vm_state: stopped {{(pid=62814) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1002.431419] env[62814]: DEBUG nova.compute.manager [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1002.487872] env[62814]: DEBUG nova.network.neutron [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Successfully created port: d2d2c877-d31a-4a2d-8e91-79d226b6a6b6 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1002.611317] env[62814]: DEBUG nova.scheduler.client.report [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1002.689075] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294177, 'name': Rename_Task, 'duration_secs': 0.203999} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.689605] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1002.689767] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-783e6d0d-d1c1-4fbe-b929-f67bc9bed819 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.697577] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1002.697577] env[62814]: value = "task-4294178" [ 1002.697577] env[62814]: _type = "Task" [ 1002.697577] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.709620] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294178, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.732451] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1002.732451] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1002.732451] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1002.732451] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1002.733477] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1002.736570] env[62814]: INFO nova.compute.manager [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Terminating instance [ 1002.926223] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1002.926513] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1002.926668] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1002.926865] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1002.927060] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1002.927303] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1002.927460] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1002.927792] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1002.927869] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1002.932276] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1002.933369] env[62814]: DEBUG nova.virt.hardware [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1002.944014] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0005f7c-7399-4e0c-ae0c-96faad314c00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.959778] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20f9527-1c12-4883-b791-623cd130a2b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.972741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1003.120935] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.735s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1003.121286] env[62814]: DEBUG nova.compute.manager [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1003.124193] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.612s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1003.124438] env[62814]: DEBUG nova.objects.instance [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lazy-loading 'resources' on Instance uuid d366a755-49b4-427b-8564-d8572a7fbbb7 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1003.212180] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294178, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.243410] env[62814]: DEBUG nova.compute.manager [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1003.243733] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1003.244673] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b0b17a-33b4-4f99-8789-007c0680f326 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.254087] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1003.255455] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8882e28c-c75a-4195-83da-0812284e379b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.264898] env[62814]: DEBUG oslo_vmware.api [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1003.264898] env[62814]: value = "task-4294179" [ 1003.264898] env[62814]: _type = "Task" [ 1003.264898] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.277748] env[62814]: DEBUG oslo_vmware.api [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294179, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.630623] env[62814]: DEBUG nova.compute.utils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1003.635331] env[62814]: DEBUG nova.compute.manager [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1003.637179] env[62814]: DEBUG nova.network.neutron [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1003.717315] env[62814]: DEBUG oslo_vmware.api [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294178, 'name': PowerOnVM_Task, 'duration_secs': 0.737565} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.720536] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1003.720863] env[62814]: INFO nova.compute.manager [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Took 14.89 seconds to spawn the instance on the hypervisor. [ 1003.721069] env[62814]: DEBUG nova.compute.manager [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1003.722731] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-510c61e9-a249-489e-85ae-32adbe79aab9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.760844] env[62814]: DEBUG nova.policy [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7bec8437999a427e865f4c6723dec2c5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8b042739d3b4927a28b9729ec5dc08b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1003.785941] env[62814]: DEBUG oslo_vmware.api [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294179, 'name': PowerOffVM_Task, 'duration_secs': 0.315991} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.785941] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1003.785941] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1003.786155] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9062efb-b37f-42cd-94c1-252d3f129195 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.857876] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1003.857876] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1003.857876] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleting the datastore file [datastore2] b5ddd05f-2027-4edf-84c5-0d2f537a95e4 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.857876] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-829c7d2d-e0ff-43b7-bc1a-ca9fd3513389 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.865658] env[62814]: DEBUG oslo_vmware.api [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1003.865658] env[62814]: value = "task-4294181" [ 1003.865658] env[62814]: _type = "Task" [ 1003.865658] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.877587] env[62814]: DEBUG oslo_vmware.api [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294181, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.980335] env[62814]: DEBUG nova.compute.manager [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1003.981821] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a849c4-369f-478e-a6c6-ab9898fc8987 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.140089] env[62814]: DEBUG nova.compute.manager [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1004.216052] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade70fe4-d98a-428c-b26e-f4aaa005358f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.224247] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a067d1-3ca2-4076-8c7a-f53568edb1fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.263760] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc9a2f2-52fb-4144-b132-0cef01716d75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.269174] env[62814]: INFO nova.compute.manager [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Took 47.26 seconds to build instance. [ 1004.275090] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9dfa5b-e8f2-4333-bab3-a9534ca1c137 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.293374] env[62814]: DEBUG nova.compute.provider_tree [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.380059] env[62814]: DEBUG oslo_vmware.api [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294181, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216673} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.380059] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.380059] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1004.380059] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1004.380059] env[62814]: INFO nova.compute.manager [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1004.380059] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1004.380059] env[62814]: DEBUG nova.compute.manager [-] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1004.380059] env[62814]: DEBUG nova.network.neutron [-] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1004.503673] env[62814]: INFO nova.compute.manager [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] instance snapshotting [ 1004.508363] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9abc77-fd42-4399-a0f6-b6ec5cd63084 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.553937] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59155587-20fe-4914-ad15-4facede144dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.629520] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "a07b964f-fe88-4c88-b6f4-61ed0973716c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1004.631490] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "a07b964f-fe88-4c88-b6f4-61ed0973716c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1004.771566] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e9fa8e8-ef78-4032-b7ec-8a0d7a2fc004 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.776s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1004.798645] env[62814]: DEBUG nova.scheduler.client.report [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1004.844209] env[62814]: DEBUG nova.network.neutron [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Successfully created port: 9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1005.048407] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1005.048407] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c6f42a01-945d-4473-91f8-fa21e886e903 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.060223] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 1005.060223] env[62814]: value = "task-4294185" [ 1005.060223] env[62814]: _type = "Task" [ 1005.060223] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.068021] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294185, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.093109] env[62814]: DEBUG nova.compute.manager [req-7560066f-7097-424e-b0f7-bbe36340c8bf req-2d84827f-69ff-4066-8628-bcfb71446217 service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Received event network-vif-plugged-d2d2c877-d31a-4a2d-8e91-79d226b6a6b6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1005.093327] env[62814]: DEBUG oslo_concurrency.lockutils [req-7560066f-7097-424e-b0f7-bbe36340c8bf req-2d84827f-69ff-4066-8628-bcfb71446217 service nova] Acquiring lock "a68901a9-9a9b-4127-bca9-64f98dfb151f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.093856] env[62814]: DEBUG oslo_concurrency.lockutils [req-7560066f-7097-424e-b0f7-bbe36340c8bf req-2d84827f-69ff-4066-8628-bcfb71446217 service nova] Lock "a68901a9-9a9b-4127-bca9-64f98dfb151f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.094273] env[62814]: DEBUG oslo_concurrency.lockutils [req-7560066f-7097-424e-b0f7-bbe36340c8bf req-2d84827f-69ff-4066-8628-bcfb71446217 service nova] Lock "a68901a9-9a9b-4127-bca9-64f98dfb151f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.094542] env[62814]: DEBUG nova.compute.manager [req-7560066f-7097-424e-b0f7-bbe36340c8bf req-2d84827f-69ff-4066-8628-bcfb71446217 service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] No waiting events found dispatching network-vif-plugged-d2d2c877-d31a-4a2d-8e91-79d226b6a6b6 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1005.094642] env[62814]: WARNING nova.compute.manager [req-7560066f-7097-424e-b0f7-bbe36340c8bf req-2d84827f-69ff-4066-8628-bcfb71446217 service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Received unexpected event network-vif-plugged-d2d2c877-d31a-4a2d-8e91-79d226b6a6b6 for instance with vm_state building and task_state spawning. [ 1005.111116] env[62814]: DEBUG nova.compute.manager [req-9e85a918-5243-4fad-9047-5ff9f991492d req-fc59ddf0-d318-43ac-9450-ba130c65d4b8 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Received event network-vif-deleted-260ddd77-4e00-4b1e-9299-fa71a7d23a57 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1005.111369] env[62814]: INFO nova.compute.manager [req-9e85a918-5243-4fad-9047-5ff9f991492d req-fc59ddf0-d318-43ac-9450-ba130c65d4b8 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Neutron deleted interface 260ddd77-4e00-4b1e-9299-fa71a7d23a57; detaching it from the instance and deleting it from the info cache [ 1005.111664] env[62814]: DEBUG nova.network.neutron [req-9e85a918-5243-4fad-9047-5ff9f991492d req-fc59ddf0-d318-43ac-9450-ba130c65d4b8 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.134991] env[62814]: DEBUG nova.compute.manager [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1005.152584] env[62814]: DEBUG nova.compute.manager [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1005.187225] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1005.187225] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1005.187225] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1005.187225] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1005.187225] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1005.187225] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1005.187225] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1005.187225] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1005.187225] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1005.188195] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1005.188533] env[62814]: DEBUG nova.virt.hardware [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1005.189956] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e44e6b-bd2b-4886-a9c2-21bf4db137ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.203299] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a18151-3f99-46ff-8f34-2654902efc9b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.307077] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.180s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.307077] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.325s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1005.307077] env[62814]: DEBUG nova.objects.instance [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lazy-loading 'resources' on Instance uuid 3e4479a0-8edd-4b37-8cc9-2c91275b88ee {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.328347] env[62814]: INFO nova.scheduler.client.report [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Deleted allocations for instance d366a755-49b4-427b-8564-d8572a7fbbb7 [ 1005.410760] env[62814]: DEBUG nova.network.neutron [-] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.498173] env[62814]: DEBUG nova.network.neutron [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Successfully updated port: d2d2c877-d31a-4a2d-8e91-79d226b6a6b6 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1005.569597] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294185, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.615739] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-481cfdf7-2083-4042-933e-bad6e87a794b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.626222] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7b3e2e-2447-4e0b-8aaf-e178fc350cb0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.675384] env[62814]: DEBUG nova.compute.manager [req-9e85a918-5243-4fad-9047-5ff9f991492d req-fc59ddf0-d318-43ac-9450-ba130c65d4b8 service nova] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Detach interface failed, port_id=260ddd77-4e00-4b1e-9299-fa71a7d23a57, reason: Instance b5ddd05f-2027-4edf-84c5-0d2f537a95e4 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1005.679923] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1005.840370] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6ae6dd65-8b88-4160-9e1d-8910aee51da3 tempest-AttachInterfacesUnderV243Test-122210080 tempest-AttachInterfacesUnderV243Test-122210080-project-member] Lock "d366a755-49b4-427b-8564-d8572a7fbbb7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.648s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1005.915124] env[62814]: INFO nova.compute.manager [-] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Took 1.54 seconds to deallocate network for instance. [ 1006.009614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquiring lock "refresh_cache-a68901a9-9a9b-4127-bca9-64f98dfb151f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.009614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquired lock "refresh_cache-a68901a9-9a9b-4127-bca9-64f98dfb151f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1006.009850] env[62814]: DEBUG nova.network.neutron [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.069994] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294185, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.331137] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30ca51b-c333-4475-8ef5-075a3e94d50a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.343221] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquiring lock "88ec5aba-f38d-4c75-af29-e3df3aa49640" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1006.344491] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lock "88ec5aba-f38d-4c75-af29-e3df3aa49640" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1006.347140] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd3e335-1bab-41bb-b27e-093b7367386c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.386883] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899c2f62-72b6-4062-8491-c904516a49f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.395070] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c55418-b600-4e5e-849c-e3e0435f272f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.411738] env[62814]: DEBUG nova.compute.provider_tree [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.427877] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1006.570644] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294185, 'name': CreateSnapshot_Task, 'duration_secs': 1.27197} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.570969] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1006.571809] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63e3fa5-1c2f-4c99-bd9f-c53199e2f596 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.588421] env[62814]: DEBUG nova.network.neutron [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1006.846842] env[62814]: DEBUG nova.compute.manager [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1006.905943] env[62814]: DEBUG nova.network.neutron [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Updating instance_info_cache with network_info: [{"id": "d2d2c877-d31a-4a2d-8e91-79d226b6a6b6", "address": "fa:16:3e:b1:1f:2b", "network": {"id": "faa3eb55-0163-456d-afd8-0e42e2fa3918", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1367444583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c3ae0fcddbe4ee7913d590a06716163", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2d2c877-d3", "ovs_interfaceid": "d2d2c877-d31a-4a2d-8e91-79d226b6a6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.914838] env[62814]: DEBUG nova.scheduler.client.report [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1007.095353] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1007.095842] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cb19eb8e-160a-4643-9f24-4696117b339d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.105497] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 1007.105497] env[62814]: value = "task-4294186" [ 1007.105497] env[62814]: _type = "Task" [ 1007.105497] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.114733] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294186, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.145368] env[62814]: DEBUG nova.compute.manager [req-0aeb5e1e-3b41-4166-bcf8-a0211c93fde2 req-fcfac76e-5929-4f11-a7ba-22446991f2e2 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Received event network-changed-75b18381-f5ef-4ce2-ab71-634c787eaffe {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1007.145731] env[62814]: DEBUG nova.compute.manager [req-0aeb5e1e-3b41-4166-bcf8-a0211c93fde2 req-fcfac76e-5929-4f11-a7ba-22446991f2e2 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Refreshing instance network info cache due to event network-changed-75b18381-f5ef-4ce2-ab71-634c787eaffe. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1007.145806] env[62814]: DEBUG oslo_concurrency.lockutils [req-0aeb5e1e-3b41-4166-bcf8-a0211c93fde2 req-fcfac76e-5929-4f11-a7ba-22446991f2e2 service nova] Acquiring lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.145964] env[62814]: DEBUG oslo_concurrency.lockutils [req-0aeb5e1e-3b41-4166-bcf8-a0211c93fde2 req-fcfac76e-5929-4f11-a7ba-22446991f2e2 service nova] Acquired lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1007.146535] env[62814]: DEBUG nova.network.neutron [req-0aeb5e1e-3b41-4166-bcf8-a0211c93fde2 req-fcfac76e-5929-4f11-a7ba-22446991f2e2 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Refreshing network info cache for port 75b18381-f5ef-4ce2-ab71-634c787eaffe {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1007.375447] env[62814]: DEBUG nova.compute.manager [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Received event network-changed-d2d2c877-d31a-4a2d-8e91-79d226b6a6b6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1007.375893] env[62814]: DEBUG nova.compute.manager [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Refreshing instance network info cache due to event network-changed-d2d2c877-d31a-4a2d-8e91-79d226b6a6b6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1007.376277] env[62814]: DEBUG oslo_concurrency.lockutils [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] Acquiring lock "refresh_cache-a68901a9-9a9b-4127-bca9-64f98dfb151f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.378667] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1007.408682] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Releasing lock "refresh_cache-a68901a9-9a9b-4127-bca9-64f98dfb151f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1007.409521] env[62814]: DEBUG nova.compute.manager [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Instance network_info: |[{"id": "d2d2c877-d31a-4a2d-8e91-79d226b6a6b6", "address": "fa:16:3e:b1:1f:2b", "network": {"id": "faa3eb55-0163-456d-afd8-0e42e2fa3918", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1367444583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c3ae0fcddbe4ee7913d590a06716163", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2d2c877-d3", "ovs_interfaceid": "d2d2c877-d31a-4a2d-8e91-79d226b6a6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1007.411131] env[62814]: DEBUG oslo_concurrency.lockutils [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] Acquired lock "refresh_cache-a68901a9-9a9b-4127-bca9-64f98dfb151f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1007.411131] env[62814]: DEBUG nova.network.neutron [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Refreshing network info cache for port d2d2c877-d31a-4a2d-8e91-79d226b6a6b6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1007.413847] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b1:1f:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56136ef6-99d7-4562-9a9f-d66fec951c5c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd2d2c877-d31a-4a2d-8e91-79d226b6a6b6', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1007.424306] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Creating folder: Project (2c3ae0fcddbe4ee7913d590a06716163). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1007.424795] env[62814]: DEBUG nova.network.neutron [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Successfully updated port: 9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.426768] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.120s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1007.429199] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b84faada-cdb8-4dc8-8134-3639926af111 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.433406] env[62814]: DEBUG oslo_concurrency.lockutils [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.597s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1007.433406] env[62814]: DEBUG nova.objects.instance [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'resources' on Instance uuid 98ba2786-023f-4d36-bdd0-e38bdd23d73c {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.446158] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Created folder: Project (2c3ae0fcddbe4ee7913d590a06716163) in parent group-v845547. [ 1007.447100] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Creating folder: Instances. Parent ref: group-v845780. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1007.448062] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d45c68c-8250-4e85-928a-ec38f837bd7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.458984] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Created folder: Instances in parent group-v845780. [ 1007.460134] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1007.460134] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1007.460134] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f855cae6-5738-430b-b26a-e617c087770b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.482259] env[62814]: INFO nova.scheduler.client.report [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Deleted allocations for instance 3e4479a0-8edd-4b37-8cc9-2c91275b88ee [ 1007.489668] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1007.489668] env[62814]: value = "task-4294189" [ 1007.489668] env[62814]: _type = "Task" [ 1007.489668] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.498616] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294189, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.625038] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294186, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.934256] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "refresh_cache-e727eeb1-c5d0-4591-80bb-31746bf976a2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.934550] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "refresh_cache-e727eeb1-c5d0-4591-80bb-31746bf976a2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1007.934590] env[62814]: DEBUG nova.network.neutron [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1007.996318] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3742739b-a8d0-40a6-829d-29f7afb57c1b tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "3e4479a0-8edd-4b37-8cc9-2c91275b88ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.721s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1008.010453] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294189, 'name': CreateVM_Task, 'duration_secs': 0.479903} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.010453] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1008.010453] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.010453] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1008.010453] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1008.010453] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff42e66e-64a1-4f02-879f-655cad14940d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.011855] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for the task: (returnval){ [ 1008.011855] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5283619c-44e4-5c81-49aa-7926de33be80" [ 1008.011855] env[62814]: _type = "Task" [ 1008.011855] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.031150] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5283619c-44e4-5c81-49aa-7926de33be80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.120920] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294186, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.310229] env[62814]: DEBUG nova.network.neutron [req-0aeb5e1e-3b41-4166-bcf8-a0211c93fde2 req-fcfac76e-5929-4f11-a7ba-22446991f2e2 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updated VIF entry in instance network info cache for port 75b18381-f5ef-4ce2-ab71-634c787eaffe. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1008.310229] env[62814]: DEBUG nova.network.neutron [req-0aeb5e1e-3b41-4166-bcf8-a0211c93fde2 req-fcfac76e-5929-4f11-a7ba-22446991f2e2 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updating instance_info_cache with network_info: [{"id": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "address": "fa:16:3e:ee:a8:e0", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b18381-f5", "ovs_interfaceid": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.410944] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca8bbfb-14d9-4fee-be74-823a56ddbe7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.420980] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15e9648-8b15-40fb-9dd8-9de79f9102e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.455925] env[62814]: DEBUG nova.network.neutron [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Updated VIF entry in instance network info cache for port d2d2c877-d31a-4a2d-8e91-79d226b6a6b6. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1008.456319] env[62814]: DEBUG nova.network.neutron [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Updating instance_info_cache with network_info: [{"id": "d2d2c877-d31a-4a2d-8e91-79d226b6a6b6", "address": "fa:16:3e:b1:1f:2b", "network": {"id": "faa3eb55-0163-456d-afd8-0e42e2fa3918", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1367444583-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c3ae0fcddbe4ee7913d590a06716163", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56136ef6-99d7-4562-9a9f-d66fec951c5c", "external-id": "nsx-vlan-transportzone-32", "segmentation_id": 32, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd2d2c877-d3", "ovs_interfaceid": "d2d2c877-d31a-4a2d-8e91-79d226b6a6b6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.458185] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c17cffe2-62ab-474f-9cc4-0167a52f5a7e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.467620] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e28fc8da-c42a-41e5-82b4-c9a0ac437850 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.484445] env[62814]: DEBUG nova.compute.provider_tree [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1008.530530] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5283619c-44e4-5c81-49aa-7926de33be80, 'name': SearchDatastore_Task, 'duration_secs': 0.015906} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.530879] env[62814]: DEBUG nova.network.neutron [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.533409] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1008.533590] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1008.533811] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.533959] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1008.534150] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1008.534433] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9fb57338-54ba-49db-a454-fd5da751e77c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.546953] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1008.549439] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1008.549439] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40a17a48-3287-4e82-a51d-b51284651de5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.555532] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for the task: (returnval){ [ 1008.555532] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521c92c9-3bf1-ab5a-c4ae-2f3c70a72984" [ 1008.555532] env[62814]: _type = "Task" [ 1008.555532] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.568649] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521c92c9-3bf1-ab5a-c4ae-2f3c70a72984, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.621114] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294186, 'name': CloneVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.817610] env[62814]: DEBUG oslo_concurrency.lockutils [req-0aeb5e1e-3b41-4166-bcf8-a0211c93fde2 req-fcfac76e-5929-4f11-a7ba-22446991f2e2 service nova] Releasing lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1008.920535] env[62814]: DEBUG nova.network.neutron [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Updating instance_info_cache with network_info: [{"id": "9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb", "address": "fa:16:3e:45:19:1e", "network": {"id": "dbc702b4-6191-4068-a1d6-4c03cc349962", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1378843546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d8b042739d3b4927a28b9729ec5dc08b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f8d050a-de", "ovs_interfaceid": "9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.966728] env[62814]: DEBUG oslo_concurrency.lockutils [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] Releasing lock "refresh_cache-a68901a9-9a9b-4127-bca9-64f98dfb151f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1008.966728] env[62814]: DEBUG nova.compute.manager [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Received event network-vif-plugged-9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1008.966728] env[62814]: DEBUG oslo_concurrency.lockutils [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] Acquiring lock "e727eeb1-c5d0-4591-80bb-31746bf976a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1008.966728] env[62814]: DEBUG oslo_concurrency.lockutils [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] Lock "e727eeb1-c5d0-4591-80bb-31746bf976a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1008.966728] env[62814]: DEBUG oslo_concurrency.lockutils [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] Lock "e727eeb1-c5d0-4591-80bb-31746bf976a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1008.967150] env[62814]: DEBUG nova.compute.manager [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] No waiting events found dispatching network-vif-plugged-9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1008.967150] env[62814]: WARNING nova.compute.manager [req-b3fed048-6b6a-47b1-8eb0-26dc571bea30 req-fb3197ee-11f2-45dc-998f-0cf448d20925 service nova] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Received unexpected event network-vif-plugged-9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb for instance with vm_state building and task_state spawning. [ 1009.011951] env[62814]: ERROR nova.scheduler.client.report [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [req-c9f91965-2571-4d40-8e58-73d9a9a6e044] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c9f91965-2571-4d40-8e58-73d9a9a6e044"}]} [ 1009.033484] env[62814]: DEBUG nova.scheduler.client.report [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1009.053755] env[62814]: DEBUG nova.scheduler.client.report [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1009.054033] env[62814]: DEBUG nova.compute.provider_tree [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1009.073157] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521c92c9-3bf1-ab5a-c4ae-2f3c70a72984, 'name': SearchDatastore_Task, 'duration_secs': 0.020499} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.074872] env[62814]: DEBUG nova.scheduler.client.report [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1009.077020] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae0d98be-6006-4bba-a69a-4e3f328dfb3d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.085689] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for the task: (returnval){ [ 1009.085689] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52632e1e-6530-a2fe-d72d-547ebbf53c2d" [ 1009.085689] env[62814]: _type = "Task" [ 1009.085689] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.096338] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52632e1e-6530-a2fe-d72d-547ebbf53c2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.113149] env[62814]: DEBUG nova.scheduler.client.report [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1009.122240] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294186, 'name': CloneVM_Task, 'duration_secs': 1.675402} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.122612] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Created linked-clone VM from snapshot [ 1009.123495] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2a6dde-f8b4-49a4-bbff-9b620b8009aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.132496] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Uploading image c8e4e708-c78f-4ee3-bc95-20830af0987b {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1009.155891] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1009.155891] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6573ef70-b268-43f9-8b9f-f452051a5a6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.163177] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 1009.163177] env[62814]: value = "task-4294191" [ 1009.163177] env[62814]: _type = "Task" [ 1009.163177] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.172378] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294191, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.418233] env[62814]: DEBUG nova.compute.manager [req-15c9eb82-97f9-4ce7-9aef-f90ffe65ce4f req-b64a012f-e2e0-48f0-8972-12cdeb216f86 service nova] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Received event network-changed-9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1009.418233] env[62814]: DEBUG nova.compute.manager [req-15c9eb82-97f9-4ce7-9aef-f90ffe65ce4f req-b64a012f-e2e0-48f0-8972-12cdeb216f86 service nova] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Refreshing instance network info cache due to event network-changed-9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1009.418233] env[62814]: DEBUG oslo_concurrency.lockutils [req-15c9eb82-97f9-4ce7-9aef-f90ffe65ce4f req-b64a012f-e2e0-48f0-8972-12cdeb216f86 service nova] Acquiring lock "refresh_cache-e727eeb1-c5d0-4591-80bb-31746bf976a2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.428768] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "refresh_cache-e727eeb1-c5d0-4591-80bb-31746bf976a2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1009.429441] env[62814]: DEBUG nova.compute.manager [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Instance network_info: |[{"id": "9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb", "address": "fa:16:3e:45:19:1e", "network": {"id": "dbc702b4-6191-4068-a1d6-4c03cc349962", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1378843546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d8b042739d3b4927a28b9729ec5dc08b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f8d050a-de", "ovs_interfaceid": "9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1009.430957] env[62814]: DEBUG oslo_concurrency.lockutils [req-15c9eb82-97f9-4ce7-9aef-f90ffe65ce4f req-b64a012f-e2e0-48f0-8972-12cdeb216f86 service nova] Acquired lock "refresh_cache-e727eeb1-c5d0-4591-80bb-31746bf976a2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1009.431177] env[62814]: DEBUG nova.network.neutron [req-15c9eb82-97f9-4ce7-9aef-f90ffe65ce4f req-b64a012f-e2e0-48f0-8972-12cdeb216f86 service nova] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Refreshing network info cache for port 9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.432583] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:19:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e6f11c0d-c73a-47f5-b02e-47bff48da0e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.442244] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Creating folder: Project (d8b042739d3b4927a28b9729ec5dc08b). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1009.445039] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bfae570a-aafc-41de-ad94-d2eff437c56a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.457914] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Created folder: Project (d8b042739d3b4927a28b9729ec5dc08b) in parent group-v845547. [ 1009.458141] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Creating folder: Instances. Parent ref: group-v845783. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1009.458406] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-90063bf7-eb63-4361-9cd2-1acb54ce84f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.471909] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Created folder: Instances in parent group-v845783. [ 1009.471909] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1009.472265] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1009.472622] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c4f7d2f-660a-415d-86cf-1930f81433c1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.502128] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.502128] env[62814]: value = "task-4294195" [ 1009.502128] env[62814]: _type = "Task" [ 1009.502128] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.510637] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294195, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.603301] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52632e1e-6530-a2fe-d72d-547ebbf53c2d, 'name': SearchDatastore_Task, 'duration_secs': 0.017075} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.603719] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1009.604096] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a68901a9-9a9b-4127-bca9-64f98dfb151f/a68901a9-9a9b-4127-bca9-64f98dfb151f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1009.604404] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-710f2400-c04e-4d08-812a-1b64f824ca07 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.617282] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for the task: (returnval){ [ 1009.617282] env[62814]: value = "task-4294196" [ 1009.617282] env[62814]: _type = "Task" [ 1009.617282] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.627620] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294196, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.676941] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294191, 'name': Destroy_Task, 'duration_secs': 0.40784} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.677100] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Destroyed the VM [ 1009.677360] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1009.677667] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b2c697e2-2a95-4ecf-8e05-a1ca9071ca88 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.684851] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 1009.684851] env[62814]: value = "task-4294197" [ 1009.684851] env[62814]: _type = "Task" [ 1009.684851] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.693773] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294197, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.739145] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cf508e-14bc-498c-a7a9-32738f3f6322 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.748579] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b48b48-9d91-4617-822b-c175d4bf9261 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.784233] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227b3ddc-a0dd-454d-8a9d-87ac92209fa4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.792715] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113614da-5645-4eae-823f-fbbb3a3fcf62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.807909] env[62814]: DEBUG nova.compute.provider_tree [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1010.018158] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294195, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.131431] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294196, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.202069] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294197, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.290596] env[62814]: DEBUG nova.network.neutron [req-15c9eb82-97f9-4ce7-9aef-f90ffe65ce4f req-b64a012f-e2e0-48f0-8972-12cdeb216f86 service nova] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Updated VIF entry in instance network info cache for port 9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1010.291067] env[62814]: DEBUG nova.network.neutron [req-15c9eb82-97f9-4ce7-9aef-f90ffe65ce4f req-b64a012f-e2e0-48f0-8972-12cdeb216f86 service nova] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Updating instance_info_cache with network_info: [{"id": "9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb", "address": "fa:16:3e:45:19:1e", "network": {"id": "dbc702b4-6191-4068-a1d6-4c03cc349962", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1378843546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d8b042739d3b4927a28b9729ec5dc08b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f8d050a-de", "ovs_interfaceid": "9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.350271] env[62814]: DEBUG nova.scheduler.client.report [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1010.350557] env[62814]: DEBUG nova.compute.provider_tree [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 119 to 120 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1010.350741] env[62814]: DEBUG nova.compute.provider_tree [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1010.517296] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294195, 'name': CreateVM_Task, 'duration_secs': 0.581128} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.517296] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1010.517936] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.518078] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1010.518418] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1010.518679] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b37535c0-e73c-4551-8228-e949c5f6cffa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.524761] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1010.524761] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d228b7-ac2d-cdad-dd6a-f7cd4f540119" [ 1010.524761] env[62814]: _type = "Task" [ 1010.524761] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.540143] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d228b7-ac2d-cdad-dd6a-f7cd4f540119, 'name': SearchDatastore_Task} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.540455] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1010.540694] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.540944] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.541106] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1010.541323] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.541794] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44dca439-d72f-496a-80a8-9a83e4682ba7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.551075] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.551075] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1010.551075] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64dec2cb-cd31-4618-a92f-a4f0dfd1287f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.558026] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1010.558026] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52724cfa-0fc4-5e87-edb0-28818b246c78" [ 1010.558026] env[62814]: _type = "Task" [ 1010.558026] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.568850] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52724cfa-0fc4-5e87-edb0-28818b246c78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.629804] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294196, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.705073} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.630122] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a68901a9-9a9b-4127-bca9-64f98dfb151f/a68901a9-9a9b-4127-bca9-64f98dfb151f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1010.630560] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1010.630835] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad5c14f0-c226-4dac-a57d-38141995e4b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.637310] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for the task: (returnval){ [ 1010.637310] env[62814]: value = "task-4294198" [ 1010.637310] env[62814]: _type = "Task" [ 1010.637310] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.646613] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294198, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.697109] env[62814]: DEBUG oslo_vmware.api [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294197, 'name': RemoveSnapshot_Task, 'duration_secs': 0.85603} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.697466] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1010.798684] env[62814]: DEBUG oslo_concurrency.lockutils [req-15c9eb82-97f9-4ce7-9aef-f90ffe65ce4f req-b64a012f-e2e0-48f0-8972-12cdeb216f86 service nova] Releasing lock "refresh_cache-e727eeb1-c5d0-4591-80bb-31746bf976a2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1010.862204] env[62814]: DEBUG oslo_concurrency.lockutils [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.428s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1010.872564] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.587s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.872564] env[62814]: INFO nova.compute.claims [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.886016] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1010.886789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1010.905168] env[62814]: INFO nova.scheduler.client.report [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleted allocations for instance 98ba2786-023f-4d36-bdd0-e38bdd23d73c [ 1011.067957] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52724cfa-0fc4-5e87-edb0-28818b246c78, 'name': SearchDatastore_Task, 'duration_secs': 0.01365} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.068923] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-778c0ed3-f5ce-45bf-9ce5-12570453395c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.074294] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1011.074294] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525b9b16-0060-86c8-572f-31910e20f670" [ 1011.074294] env[62814]: _type = "Task" [ 1011.074294] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.083138] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525b9b16-0060-86c8-572f-31910e20f670, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.149794] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294198, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074425} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.149794] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1011.150571] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63273b1f-47ed-4c1b-a03d-a0fe9e535d57 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.175709] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] a68901a9-9a9b-4127-bca9-64f98dfb151f/a68901a9-9a9b-4127-bca9-64f98dfb151f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.176083] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8f8282d-e812-4053-941c-357571865021 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.199277] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for the task: (returnval){ [ 1011.199277] env[62814]: value = "task-4294200" [ 1011.199277] env[62814]: _type = "Task" [ 1011.199277] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.203060] env[62814]: WARNING nova.compute.manager [None req-792de3bd-7910-4d81-858e-0246ac8f71c5 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Image not found during snapshot: nova.exception.ImageNotFound: Image c8e4e708-c78f-4ee3-bc95-20830af0987b could not be found. [ 1011.209721] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294200, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.389336] env[62814]: DEBUG nova.compute.manager [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1011.415682] env[62814]: DEBUG oslo_concurrency.lockutils [None req-55563860-3c2b-4a7d-b46e-bd3d93dcfbd8 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "98ba2786-023f-4d36-bdd0-e38bdd23d73c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.552s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.478035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.478242] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.495509] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "40e2d845-0211-4c84-aef7-94014f999e1d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.495759] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "40e2d845-0211-4c84-aef7-94014f999e1d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.495973] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "40e2d845-0211-4c84-aef7-94014f999e1d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.496266] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "40e2d845-0211-4c84-aef7-94014f999e1d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1011.496360] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "40e2d845-0211-4c84-aef7-94014f999e1d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1011.498305] env[62814]: INFO nova.compute.manager [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Terminating instance [ 1011.586723] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525b9b16-0060-86c8-572f-31910e20f670, 'name': SearchDatastore_Task, 'duration_secs': 0.011177} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.586723] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1011.586990] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] e727eeb1-c5d0-4591-80bb-31746bf976a2/e727eeb1-c5d0-4591-80bb-31746bf976a2.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1011.587156] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e0ea727a-7f0f-4588-a99f-1f83f118b64b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.594538] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1011.594538] env[62814]: value = "task-4294201" [ 1011.594538] env[62814]: _type = "Task" [ 1011.594538] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.602665] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294201, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.708999] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294200, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.922040] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.983105] env[62814]: DEBUG nova.compute.manager [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1011.988353] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1011.988687] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1012.002046] env[62814]: DEBUG nova.compute.manager [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1012.002260] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.003787] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6046b3ef-5826-4dc7-917b-91c81ccf6806 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.015201] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1012.015201] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dd7f73dd-6535-4f59-9583-4a9d11da7f96 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.027029] env[62814]: DEBUG oslo_vmware.api [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1012.027029] env[62814]: value = "task-4294202" [ 1012.027029] env[62814]: _type = "Task" [ 1012.027029] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.038333] env[62814]: DEBUG oslo_vmware.api [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294202, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.113968] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294201, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.212339] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294200, 'name': ReconfigVM_Task, 'duration_secs': 0.5174} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.212672] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Reconfigured VM instance instance-0000004e to attach disk [datastore2] a68901a9-9a9b-4127-bca9-64f98dfb151f/a68901a9-9a9b-4127-bca9-64f98dfb151f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.213335] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-235653a7-4896-45a1-8d72-87096bb8c4b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.215836] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "fd1867e3-2523-4969-a157-b14c650f3779" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1012.215836] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "fd1867e3-2523-4969-a157-b14c650f3779" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1012.215836] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "fd1867e3-2523-4969-a157-b14c650f3779-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1012.215836] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "fd1867e3-2523-4969-a157-b14c650f3779-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1012.216088] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "fd1867e3-2523-4969-a157-b14c650f3779-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1012.218494] env[62814]: INFO nova.compute.manager [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Terminating instance [ 1012.227111] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for the task: (returnval){ [ 1012.227111] env[62814]: value = "task-4294203" [ 1012.227111] env[62814]: _type = "Task" [ 1012.227111] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.237423] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294203, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.360668] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8f568e6-6f7b-47a2-ba92-2719a09107fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.368501] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68632de3-a4be-40e3-b8d3-2eb57f3c0438 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.403992] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d317e69e-aad2-4907-a869-f4d7d00728e1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.410723] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25a3e32-9194-49aa-a185-4c13bf5a3de8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.425334] env[62814]: DEBUG nova.compute.provider_tree [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1012.495808] env[62814]: DEBUG nova.compute.manager [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1012.526307] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1012.537612] env[62814]: DEBUG oslo_vmware.api [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294202, 'name': PowerOffVM_Task, 'duration_secs': 0.240171} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.537937] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1012.538171] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1012.538298] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1e4e8eb8-1241-49c3-86f1-0650f5772549 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.602531] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1012.602858] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1012.603177] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleting the datastore file [datastore2] 40e2d845-0211-4c84-aef7-94014f999e1d {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.607371] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d12e23a-041d-4e09-aabd-c1a21a05ac4a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.609948] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294201, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557057} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.610292] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] e727eeb1-c5d0-4591-80bb-31746bf976a2/e727eeb1-c5d0-4591-80bb-31746bf976a2.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1012.610595] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.611349] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b692f86d-f236-4aa1-badd-e28bb6e62056 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.615035] env[62814]: DEBUG oslo_vmware.api [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1012.615035] env[62814]: value = "task-4294205" [ 1012.615035] env[62814]: _type = "Task" [ 1012.615035] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.620336] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1012.620336] env[62814]: value = "task-4294206" [ 1012.620336] env[62814]: _type = "Task" [ 1012.620336] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.628708] env[62814]: DEBUG oslo_vmware.api [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294205, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.634173] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294206, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.727038] env[62814]: DEBUG nova.compute.manager [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1012.727704] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1012.728595] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a5cec4-f946-4a5f-9a40-881090389b88 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.742327] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294203, 'name': Rename_Task, 'duration_secs': 0.19786} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.744888] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1012.745301] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1012.745571] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a718856-2776-411c-9f52-6fa307e2cf1f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.747443] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f70e517f-1c93-4ef4-bc15-dc61cd896ea3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.755256] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for the task: (returnval){ [ 1012.755256] env[62814]: value = "task-4294207" [ 1012.755256] env[62814]: _type = "Task" [ 1012.755256] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.756750] env[62814]: DEBUG oslo_vmware.api [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 1012.756750] env[62814]: value = "task-4294208" [ 1012.756750] env[62814]: _type = "Task" [ 1012.756750] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.769046] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294207, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.772636] env[62814]: DEBUG oslo_vmware.api [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294208, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.929160] env[62814]: DEBUG nova.scheduler.client.report [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1013.026826] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.130860] env[62814]: DEBUG oslo_vmware.api [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294205, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176621} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.131336] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.131517] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1013.132303] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1013.132303] env[62814]: INFO nova.compute.manager [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1013.132659] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1013.135443] env[62814]: DEBUG nova.compute.manager [-] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1013.135563] env[62814]: DEBUG nova.network.neutron [-] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1013.137373] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294206, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0713} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.137650] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1013.141711] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fe574b-5905-443e-be12-c41bab824b31 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.160643] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] e727eeb1-c5d0-4591-80bb-31746bf976a2/e727eeb1-c5d0-4591-80bb-31746bf976a2.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1013.160971] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28db1639-2766-40bf-b3c2-6f736b422df1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.177275] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "d9242042-6209-4b04-bf00-00dd04d9d6a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.177522] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "d9242042-6209-4b04-bf00-00dd04d9d6a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.177763] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "d9242042-6209-4b04-bf00-00dd04d9d6a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.177942] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "d9242042-6209-4b04-bf00-00dd04d9d6a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.178129] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "d9242042-6209-4b04-bf00-00dd04d9d6a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.180512] env[62814]: INFO nova.compute.manager [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Terminating instance [ 1013.184627] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1013.184627] env[62814]: value = "task-4294210" [ 1013.184627] env[62814]: _type = "Task" [ 1013.184627] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.194641] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294210, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.274411] env[62814]: DEBUG oslo_vmware.api [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294208, 'name': PowerOffVM_Task, 'duration_secs': 0.199659} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.277030] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1013.277136] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1013.277379] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294207, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.277774] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da726c70-2e6d-4772-9fce-9495812183b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.341575] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1013.341833] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1013.342028] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleting the datastore file [datastore2] fd1867e3-2523-4969-a157-b14c650f3779 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1013.342305] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-411ef6cc-d125-457a-ad01-6be698dba603 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.352113] env[62814]: DEBUG oslo_vmware.api [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for the task: (returnval){ [ 1013.352113] env[62814]: value = "task-4294212" [ 1013.352113] env[62814]: _type = "Task" [ 1013.352113] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.365390] env[62814]: DEBUG oslo_vmware.api [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294212, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.438701] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.439633] env[62814]: DEBUG nova.compute.manager [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1013.442958] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 23.083s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.443075] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1013.443762] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1013.443762] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.527s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1013.443762] env[62814]: DEBUG nova.objects.instance [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lazy-loading 'resources' on Instance uuid 4ed66b36-b6c6-4673-9c03-169a01134574 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1013.445603] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5249e485-6f8c-44f7-892f-1f881d22d182 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.455416] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502ebc05-1462-4a0a-b3b2-e46d7963adc3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.474368] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f52ccb8-664d-49ee-a9da-6273abc11418 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.483804] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a21d94ac-ecb5-43e4-8d03-d2c0605100dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.523315] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=178090MB free_disk=143GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1013.523518] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1013.689673] env[62814]: DEBUG nova.compute.manager [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1013.689899] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1013.690723] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a568c6f-25d7-4229-a642-0320901d5bf8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.701656] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294210, 'name': ReconfigVM_Task, 'duration_secs': 0.374943} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.702147] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Reconfigured VM instance instance-0000004f to attach disk [datastore2] e727eeb1-c5d0-4591-80bb-31746bf976a2/e727eeb1-c5d0-4591-80bb-31746bf976a2.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.702879] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1013.703457] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19cd056f-81df-43d8-abfd-a7c8d32d4a98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.704719] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4fb5712e-0020-4337-a7a6-b4feb849d76b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.711733] env[62814]: DEBUG oslo_vmware.api [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 1013.711733] env[62814]: value = "task-4294213" [ 1013.711733] env[62814]: _type = "Task" [ 1013.711733] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.713477] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1013.713477] env[62814]: value = "task-4294214" [ 1013.713477] env[62814]: _type = "Task" [ 1013.713477] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.734385] env[62814]: DEBUG oslo_vmware.api [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294213, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.737926] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294214, 'name': Rename_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.773685] env[62814]: DEBUG oslo_vmware.api [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294207, 'name': PowerOnVM_Task, 'duration_secs': 0.818301} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.774077] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1013.774316] env[62814]: INFO nova.compute.manager [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Took 11.34 seconds to spawn the instance on the hypervisor. [ 1013.774575] env[62814]: DEBUG nova.compute.manager [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1013.775604] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d74b3f-612b-43ec-aeab-c543757e536b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.844765] env[62814]: DEBUG nova.compute.manager [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Stashing vm_state: active {{(pid=62814) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1013.867975] env[62814]: DEBUG oslo_vmware.api [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Task: {'id': task-4294212, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.216695} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.870451] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.870733] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1013.870990] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1013.871413] env[62814]: INFO nova.compute.manager [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1013.871729] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1013.872036] env[62814]: DEBUG nova.compute.manager [-] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1013.872159] env[62814]: DEBUG nova.network.neutron [-] [instance: fd1867e3-2523-4969-a157-b14c650f3779] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1013.950638] env[62814]: DEBUG nova.compute.utils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1013.955658] env[62814]: DEBUG nova.compute.manager [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1013.955862] env[62814]: DEBUG nova.network.neutron [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1014.050357] env[62814]: DEBUG nova.policy [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1038a87615ae415ea4462635cb7d391d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '869db97f30f544f49e6a0c0e54bea463', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1014.238792] env[62814]: DEBUG oslo_vmware.api [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294213, 'name': PowerOffVM_Task, 'duration_secs': 0.494058} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.238792] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294214, 'name': Rename_Task, 'duration_secs': 0.15868} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.241364] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.241364] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1014.241555] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1014.242058] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-32607926-ae12-413c-b87a-60c4c2bf6a06 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.243591] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fc4e2f99-6c45-4e54-b4e7-3bbc76a40222 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.251418] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1014.251418] env[62814]: value = "task-4294215" [ 1014.251418] env[62814]: _type = "Task" [ 1014.251418] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.262734] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294215, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.300392] env[62814]: INFO nova.compute.manager [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Took 40.87 seconds to build instance. [ 1014.334715] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1014.334839] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1014.335119] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Deleting the datastore file [datastore2] d9242042-6209-4b04-bf00-00dd04d9d6a0 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.335872] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af0700fc-4496-445e-9c13-e2f2891ff4ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.346927] env[62814]: DEBUG oslo_vmware.api [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 1014.346927] env[62814]: value = "task-4294217" [ 1014.346927] env[62814]: _type = "Task" [ 1014.346927] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.366394] env[62814]: DEBUG oslo_vmware.api [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294217, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.372793] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1014.404232] env[62814]: DEBUG nova.compute.manager [req-13964fbe-a595-4928-806c-59300dcde72f req-e0a69c5b-d74f-42f0-9c6c-efd52d9639f1 service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Received event network-vif-deleted-029acc9a-5528-4c96-a5a7-fa02e89a2bb9 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1014.404532] env[62814]: INFO nova.compute.manager [req-13964fbe-a595-4928-806c-59300dcde72f req-e0a69c5b-d74f-42f0-9c6c-efd52d9639f1 service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Neutron deleted interface 029acc9a-5528-4c96-a5a7-fa02e89a2bb9; detaching it from the instance and deleting it from the info cache [ 1014.404781] env[62814]: DEBUG nova.network.neutron [req-13964fbe-a595-4928-806c-59300dcde72f req-e0a69c5b-d74f-42f0-9c6c-efd52d9639f1 service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.410288] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-827f6f61-813d-4451-b095-7e55b3e35854 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.421708] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de4a1ec-e5f3-4b0c-9a9d-395d83f52ef4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.460708] env[62814]: DEBUG nova.compute.manager [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1014.463934] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2135b1-1c8b-459b-9535-446d2ccee5f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.482533] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbf8436-6a82-4733-a1d8-212a64c8046a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.502033] env[62814]: DEBUG nova.compute.provider_tree [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1014.586653] env[62814]: DEBUG nova.network.neutron [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Successfully created port: 9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1014.764746] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294215, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.803152] env[62814]: DEBUG oslo_concurrency.lockutils [None req-867ef93e-ec29-4a98-9002-c34f0fe23cb1 tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Lock "a68901a9-9a9b-4127-bca9-64f98dfb151f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.411s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1014.806783] env[62814]: DEBUG nova.network.neutron [-] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.863080] env[62814]: DEBUG oslo_vmware.api [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294217, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.261812} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.863355] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1014.863547] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1014.863887] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1014.864138] env[62814]: INFO nova.compute.manager [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1014.864450] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1014.864648] env[62814]: DEBUG nova.compute.manager [-] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1014.864742] env[62814]: DEBUG nova.network.neutron [-] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1014.907967] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a10b824d-caaf-45df-aab0-cc88f00c05b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.919592] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02f648f-f753-49f1-a22c-ca4ee46570aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.957249] env[62814]: DEBUG nova.compute.manager [req-13964fbe-a595-4928-806c-59300dcde72f req-e0a69c5b-d74f-42f0-9c6c-efd52d9639f1 service nova] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Detach interface failed, port_id=029acc9a-5528-4c96-a5a7-fa02e89a2bb9, reason: Instance 40e2d845-0211-4c84-aef7-94014f999e1d could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1015.009861] env[62814]: DEBUG nova.network.neutron [-] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.030804] env[62814]: ERROR nova.scheduler.client.report [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] [req-aaa2994f-ede6-4381-b9f3-b0c938cb4c19] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-aaa2994f-ede6-4381-b9f3-b0c938cb4c19"}]} [ 1015.047366] env[62814]: DEBUG nova.scheduler.client.report [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1015.064322] env[62814]: DEBUG nova.scheduler.client.report [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1015.064615] env[62814]: DEBUG nova.compute.provider_tree [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1015.076776] env[62814]: DEBUG nova.scheduler.client.report [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1015.096291] env[62814]: DEBUG nova.scheduler.client.report [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1015.263867] env[62814]: DEBUG oslo_vmware.api [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294215, 'name': PowerOnVM_Task, 'duration_secs': 0.574162} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.264192] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1015.264395] env[62814]: INFO nova.compute.manager [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Took 10.11 seconds to spawn the instance on the hypervisor. [ 1015.265538] env[62814]: DEBUG nova.compute.manager [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1015.269021] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d88ec1-a311-4247-aec1-07a46af419dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.309976] env[62814]: INFO nova.compute.manager [-] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Took 2.17 seconds to deallocate network for instance. [ 1015.479516] env[62814]: DEBUG nova.compute.manager [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1015.518300] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1015.518554] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1015.518759] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1015.518890] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1015.519074] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1015.519202] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1015.519446] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1015.519609] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1015.519777] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1015.519945] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1015.520163] env[62814]: DEBUG nova.virt.hardware [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1015.520618] env[62814]: INFO nova.compute.manager [-] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Took 1.65 seconds to deallocate network for instance. [ 1015.521466] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b8a085-75ac-4263-a540-44d21143e8be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.535515] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756c883d-865e-4fae-9021-a602ed73ed6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.566505] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c27526-3b2c-4a7a-96d1-719c196b6e19 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.576196] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a96f636-dcae-4fcf-a7a7-6e8924c5d442 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.610605] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121f156b-82ee-4481-943b-e65894b922d8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.620687] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a53573-6266-4ab3-aa4f-2d297774fdb8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.637991] env[62814]: DEBUG nova.compute.provider_tree [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1015.787821] env[62814]: DEBUG nova.network.neutron [-] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.800133] env[62814]: INFO nova.compute.manager [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Took 40.81 seconds to build instance. [ 1015.821266] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.035649] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.070569] env[62814]: DEBUG oslo_concurrency.lockutils [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquiring lock "a68901a9-9a9b-4127-bca9-64f98dfb151f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.073583] env[62814]: DEBUG oslo_concurrency.lockutils [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Lock "a68901a9-9a9b-4127-bca9-64f98dfb151f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1016.073583] env[62814]: DEBUG oslo_concurrency.lockutils [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquiring lock "a68901a9-9a9b-4127-bca9-64f98dfb151f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.073583] env[62814]: DEBUG oslo_concurrency.lockutils [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Lock "a68901a9-9a9b-4127-bca9-64f98dfb151f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1016.073583] env[62814]: DEBUG oslo_concurrency.lockutils [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Lock "a68901a9-9a9b-4127-bca9-64f98dfb151f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.075076] env[62814]: INFO nova.compute.manager [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Terminating instance [ 1016.186335] env[62814]: DEBUG nova.scheduler.client.report [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 125 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1016.186704] env[62814]: DEBUG nova.compute.provider_tree [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 125 to 126 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1016.187034] env[62814]: DEBUG nova.compute.provider_tree [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1016.297976] env[62814]: INFO nova.compute.manager [-] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Took 1.43 seconds to deallocate network for instance. [ 1016.300705] env[62814]: DEBUG nova.compute.manager [req-95cd2ba0-3873-455f-9b4c-c6708f669503 req-5c6c8ae6-d73a-45d3-aa90-c021434bf7bb service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Received event network-vif-plugged-9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1016.300705] env[62814]: DEBUG oslo_concurrency.lockutils [req-95cd2ba0-3873-455f-9b4c-c6708f669503 req-5c6c8ae6-d73a-45d3-aa90-c021434bf7bb service nova] Acquiring lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.300705] env[62814]: DEBUG oslo_concurrency.lockutils [req-95cd2ba0-3873-455f-9b4c-c6708f669503 req-5c6c8ae6-d73a-45d3-aa90-c021434bf7bb service nova] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1016.300705] env[62814]: DEBUG oslo_concurrency.lockutils [req-95cd2ba0-3873-455f-9b4c-c6708f669503 req-5c6c8ae6-d73a-45d3-aa90-c021434bf7bb service nova] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.300705] env[62814]: DEBUG nova.compute.manager [req-95cd2ba0-3873-455f-9b4c-c6708f669503 req-5c6c8ae6-d73a-45d3-aa90-c021434bf7bb service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] No waiting events found dispatching network-vif-plugged-9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1016.300705] env[62814]: WARNING nova.compute.manager [req-95cd2ba0-3873-455f-9b4c-c6708f669503 req-5c6c8ae6-d73a-45d3-aa90-c021434bf7bb service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Received unexpected event network-vif-plugged-9950de47-d55a-480b-b88e-818e09b9bc80 for instance with vm_state building and task_state spawning. [ 1016.307507] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c8497716-74a3-4edd-9784-4f6c35979721 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "e727eeb1-c5d0-4591-80bb-31746bf976a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.340s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.331800] env[62814]: DEBUG nova.network.neutron [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Successfully updated port: 9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1016.386861] env[62814]: DEBUG oslo_vmware.rw_handles [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52453c89-9711-2b9b-df20-96f0485f5086/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1016.387930] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11568636-c0d7-48f2-9368-a58d34031644 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.396447] env[62814]: DEBUG oslo_vmware.rw_handles [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52453c89-9711-2b9b-df20-96f0485f5086/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1016.396447] env[62814]: ERROR oslo_vmware.rw_handles [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52453c89-9711-2b9b-df20-96f0485f5086/disk-0.vmdk due to incomplete transfer. [ 1016.396765] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0c7691ca-07c6-4637-adce-25ae9197acde {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.406168] env[62814]: DEBUG oslo_vmware.rw_handles [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52453c89-9711-2b9b-df20-96f0485f5086/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1016.406582] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Uploaded image 83b049c5-1828-43ae-9dc4-e43a6a030542 to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1016.408952] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1016.409248] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-28581d89-880f-4ca6-9d45-8c5dbae594a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.418145] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 1016.418145] env[62814]: value = "task-4294219" [ 1016.418145] env[62814]: _type = "Task" [ 1016.418145] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.426714] env[62814]: INFO nova.compute.manager [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Rescuing [ 1016.426714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "refresh_cache-e727eeb1-c5d0-4591-80bb-31746bf976a2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.426714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "refresh_cache-e727eeb1-c5d0-4591-80bb-31746bf976a2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1016.426714] env[62814]: DEBUG nova.network.neutron [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1016.429553] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294219, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.445297] env[62814]: DEBUG nova.compute.manager [req-25ad7693-91c1-46cb-82aa-a24074158079 req-df141e45-40f9-4378-be3b-15e4aa5e223d service nova] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Received event network-vif-deleted-26bd8c43-7afd-47eb-b53a-52d4a7847cec {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1016.445567] env[62814]: DEBUG nova.compute.manager [req-25ad7693-91c1-46cb-82aa-a24074158079 req-df141e45-40f9-4378-be3b-15e4aa5e223d service nova] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Received event network-vif-deleted-a2e379d6-9cd3-4a22-85c2-72628174b4a6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1016.585647] env[62814]: DEBUG nova.compute.manager [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1016.585647] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1016.588294] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac598a2-cad7-423c-9974-275228caef82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.599531] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1016.599531] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32deaa4f-7ca8-4d99-881e-8674c602c92e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.610650] env[62814]: DEBUG oslo_vmware.api [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for the task: (returnval){ [ 1016.610650] env[62814]: value = "task-4294220" [ 1016.610650] env[62814]: _type = "Task" [ 1016.610650] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.620803] env[62814]: DEBUG oslo_vmware.api [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294220, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.695169] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.250s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1016.697485] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.440s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1016.697485] env[62814]: DEBUG nova.objects.instance [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1016.729381] env[62814]: INFO nova.scheduler.client.report [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Deleted allocations for instance 4ed66b36-b6c6-4673-9c03-169a01134574 [ 1016.807678] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1016.835931] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.836107] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1016.836268] env[62814]: DEBUG nova.network.neutron [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1016.939108] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294219, 'name': Destroy_Task, 'duration_secs': 0.450162} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.939108] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Destroyed the VM [ 1016.940358] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1016.940997] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-57bceb95-e2e4-4d04-93cc-ce4aa4fd8231 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.948639] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 1016.948639] env[62814]: value = "task-4294221" [ 1016.948639] env[62814]: _type = "Task" [ 1016.948639] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.964043] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294221, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.123591] env[62814]: DEBUG oslo_vmware.api [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294220, 'name': PowerOffVM_Task, 'duration_secs': 0.215781} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.126606] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1017.126882] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1017.127124] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-08862219-74b3-494b-afde-f80423415389 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.177847] env[62814]: DEBUG nova.network.neutron [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Updating instance_info_cache with network_info: [{"id": "9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb", "address": "fa:16:3e:45:19:1e", "network": {"id": "dbc702b4-6191-4068-a1d6-4c03cc349962", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1378843546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d8b042739d3b4927a28b9729ec5dc08b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f8d050a-de", "ovs_interfaceid": "9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.198631] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1017.198825] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1017.199071] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Deleting the datastore file [datastore2] a68901a9-9a9b-4127-bca9-64f98dfb151f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1017.199376] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05e3a6d6-19d9-42d0-989c-53737422faf7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.229239] env[62814]: DEBUG oslo_vmware.api [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for the task: (returnval){ [ 1017.229239] env[62814]: value = "task-4294223" [ 1017.229239] env[62814]: _type = "Task" [ 1017.229239] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.243089] env[62814]: DEBUG oslo_vmware.api [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294223, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.244893] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c762d70-ea5a-43dd-a430-fc93daa800e4 tempest-ServersAdminTestJSON-1687339248 tempest-ServersAdminTestJSON-1687339248-project-member] Lock "4ed66b36-b6c6-4673-9c03-169a01134574" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.593s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1017.405809] env[62814]: DEBUG nova.network.neutron [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1017.463324] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294221, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.683178] env[62814]: DEBUG oslo_concurrency.lockutils [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "refresh_cache-e727eeb1-c5d0-4591-80bb-31746bf976a2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1017.712939] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c33134f1-c6e3-4e4a-a521-8c882c6b83ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1017.714486] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.874s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1017.715976] env[62814]: INFO nova.compute.claims [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1017.729983] env[62814]: DEBUG nova.network.neutron [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updating instance_info_cache with network_info: [{"id": "9950de47-d55a-480b-b88e-818e09b9bc80", "address": "fa:16:3e:0c:71:2f", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9950de47-d5", "ovs_interfaceid": "9950de47-d55a-480b-b88e-818e09b9bc80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.750160] env[62814]: DEBUG oslo_vmware.api [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Task: {'id': task-4294223, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.288808} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.750160] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1017.750160] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1017.750160] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1017.751946] env[62814]: INFO nova.compute.manager [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1017.751946] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1017.751946] env[62814]: DEBUG nova.compute.manager [-] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1017.751946] env[62814]: DEBUG nova.network.neutron [-] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1017.963219] env[62814]: DEBUG oslo_vmware.api [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294221, 'name': RemoveSnapshot_Task, 'duration_secs': 0.584946} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.963523] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1017.963759] env[62814]: INFO nova.compute.manager [None req-88e2a796-625c-4260-8d62-c8200a137a53 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Took 21.51 seconds to snapshot the instance on the hypervisor. [ 1018.242283] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1018.242822] env[62814]: DEBUG nova.compute.manager [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Instance network_info: |[{"id": "9950de47-d55a-480b-b88e-818e09b9bc80", "address": "fa:16:3e:0c:71:2f", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9950de47-d5", "ovs_interfaceid": "9950de47-d55a-480b-b88e-818e09b9bc80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1018.243981] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:71:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9950de47-d55a-480b-b88e-818e09b9bc80', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1018.257025] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Creating folder: Project (869db97f30f544f49e6a0c0e54bea463). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1018.257025] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d973450d-1f97-4b4b-9462-895bd60ff62b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.268330] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Created folder: Project (869db97f30f544f49e6a0c0e54bea463) in parent group-v845547. [ 1018.268641] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Creating folder: Instances. Parent ref: group-v845787. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1018.268942] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bb43b2ed-e2c1-4e5b-8c7b-22cb6ec6331b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.280449] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Created folder: Instances in parent group-v845787. [ 1018.280828] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1018.282034] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1018.282034] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a86b84fe-4f7c-4b60-ab71-030e58552689 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.302437] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1018.302437] env[62814]: value = "task-4294226" [ 1018.302437] env[62814]: _type = "Task" [ 1018.302437] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.314067] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294226, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.373539] env[62814]: DEBUG nova.compute.manager [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Received event network-changed-9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1018.374025] env[62814]: DEBUG nova.compute.manager [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Refreshing instance network info cache due to event network-changed-9950de47-d55a-480b-b88e-818e09b9bc80. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1018.374379] env[62814]: DEBUG oslo_concurrency.lockutils [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] Acquiring lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.374694] env[62814]: DEBUG oslo_concurrency.lockutils [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] Acquired lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1018.375050] env[62814]: DEBUG nova.network.neutron [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Refreshing network info cache for port 9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1018.511138] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1018.511138] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1018.679612] env[62814]: DEBUG nova.network.neutron [-] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1018.817410] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294226, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.014141] env[62814]: DEBUG nova.compute.manager [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1019.181981] env[62814]: INFO nova.compute.manager [-] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Took 1.43 seconds to deallocate network for instance. [ 1019.242333] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1019.243364] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5bd57253-097c-4889-9fed-8893cc683080 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.254337] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1019.254337] env[62814]: value = "task-4294227" [ 1019.254337] env[62814]: _type = "Task" [ 1019.254337] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.271244] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294227, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.285110] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d401d10a-b399-4446-ace8-035634c0a30a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.293980] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf88bfd0-c5f8-41b4-b8c5-44a6867c8dd0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.336031] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2342e986-f10e-4412-9bac-266f503b51b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.349956] env[62814]: DEBUG nova.network.neutron [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updated VIF entry in instance network info cache for port 9950de47-d55a-480b-b88e-818e09b9bc80. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1019.349956] env[62814]: DEBUG nova.network.neutron [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updating instance_info_cache with network_info: [{"id": "9950de47-d55a-480b-b88e-818e09b9bc80", "address": "fa:16:3e:0c:71:2f", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9950de47-d5", "ovs_interfaceid": "9950de47-d55a-480b-b88e-818e09b9bc80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.357945] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294226, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.359294] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ebb6a0-53a7-4a33-a0df-bb59c7a82bb3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.377731] env[62814]: DEBUG nova.compute.provider_tree [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.542324] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.691822] env[62814]: DEBUG oslo_concurrency.lockutils [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1019.769395] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294227, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.846021] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294226, 'name': CreateVM_Task, 'duration_secs': 1.481331} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.846021] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1019.846021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.846021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1019.846021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1019.846021] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7defb6b1-390f-4550-812b-2f3efc68ec13 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.857021] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1019.857021] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52452ca9-97dc-813e-312b-a61a71dfcbcf" [ 1019.857021] env[62814]: _type = "Task" [ 1019.857021] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.858307] env[62814]: DEBUG oslo_concurrency.lockutils [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] Releasing lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.859192] env[62814]: DEBUG nova.compute.manager [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Received event network-vif-deleted-d2d2c877-d31a-4a2d-8e91-79d226b6a6b6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1019.859535] env[62814]: INFO nova.compute.manager [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Neutron deleted interface d2d2c877-d31a-4a2d-8e91-79d226b6a6b6; detaching it from the instance and deleting it from the info cache [ 1019.859839] env[62814]: DEBUG nova.network.neutron [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.870154] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52452ca9-97dc-813e-312b-a61a71dfcbcf, 'name': SearchDatastore_Task, 'duration_secs': 0.013591} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.870992] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.871394] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1019.871749] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.872013] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1019.872308] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1019.877462] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c5231921-91fd-4c6b-9382-dc283845f6ef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.883021] env[62814]: DEBUG nova.scheduler.client.report [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1019.886784] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1019.887265] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1019.888235] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bba9632-0cae-4102-a14e-bd568f9b42d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.898346] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1019.898346] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f8eb3e-195f-2b31-bcb0-7273828500ba" [ 1019.898346] env[62814]: _type = "Task" [ 1019.898346] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.914398] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f8eb3e-195f-2b31-bcb0-7273828500ba, 'name': SearchDatastore_Task, 'duration_secs': 0.011144} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.915541] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3df3086-ae8a-4a49-94f6-e6ae8b2fb5e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.923135] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1019.923135] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521cdb80-07df-1dde-6463-5fd21c530b44" [ 1019.923135] env[62814]: _type = "Task" [ 1019.923135] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.936025] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521cdb80-07df-1dde-6463-5fd21c530b44, 'name': SearchDatastore_Task, 'duration_secs': 0.010635} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.936285] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1019.936545] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] edcdb1e5-ed9d-49e6-97e6-ea7629682547/edcdb1e5-ed9d-49e6-97e6-ea7629682547.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1019.936807] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32b6234b-846b-4343-859d-4e65ea352191 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.945066] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1019.945066] env[62814]: value = "task-4294229" [ 1019.945066] env[62814]: _type = "Task" [ 1019.945066] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.959397] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294229, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.271167] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294227, 'name': PowerOffVM_Task, 'duration_secs': 0.595374} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.271411] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1020.272338] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f94d46-448e-4f7d-96ff-1058cac16640 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.294456] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b436c5-bde0-45cb-879b-b7be2d1a9613 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.333658] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1020.333998] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bc39cc5-b2ec-4292-929d-ed406c99656f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.344931] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1020.344931] env[62814]: value = "task-4294230" [ 1020.344931] env[62814]: _type = "Task" [ 1020.344931] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.355793] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1020.356039] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1020.356319] env[62814]: DEBUG oslo_concurrency.lockutils [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1020.356553] env[62814]: DEBUG oslo_concurrency.lockutils [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1020.356675] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1020.356943] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7ade2187-85c1-4746-aadd-a74b23685053 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.362995] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41198452-9fb1-4ed8-a98e-bc95b6e3f28c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.379027] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75981d06-ca97-4ad3-9acd-2fe98890d043 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.387199] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1020.387455] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1020.388840] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.674s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1020.389401] env[62814]: DEBUG nova.compute.manager [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1020.392067] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8034779-76a1-45f9-ae53-40490e522381 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.403265] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.459s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1020.405253] env[62814]: INFO nova.compute.claims [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1020.418672] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1020.418672] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529cc526-219f-7895-583e-e716c6dec93f" [ 1020.418672] env[62814]: _type = "Task" [ 1020.418672] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.446878] env[62814]: DEBUG nova.compute.manager [req-096826b1-152f-4a99-9c1c-892a41feed1f req-a67502ab-1b9b-43df-9922-24a21dfe901d service nova] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Detach interface failed, port_id=d2d2c877-d31a-4a2d-8e91-79d226b6a6b6, reason: Instance a68901a9-9a9b-4127-bca9-64f98dfb151f could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1020.455107] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529cc526-219f-7895-583e-e716c6dec93f, 'name': SearchDatastore_Task, 'duration_secs': 0.032695} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.459057] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294229, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.459348] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afd91465-5260-4c7e-a613-e593ab6b2988 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.465791] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1020.465791] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b636e2-ac8d-9abe-5fc3-87b332f6c46c" [ 1020.465791] env[62814]: _type = "Task" [ 1020.465791] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.475496] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b636e2-ac8d-9abe-5fc3-87b332f6c46c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.903978] env[62814]: DEBUG nova.compute.utils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1020.909896] env[62814]: DEBUG nova.compute.manager [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1020.911066] env[62814]: DEBUG nova.network.neutron [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1020.957193] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294229, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51686} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.957504] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] edcdb1e5-ed9d-49e6-97e6-ea7629682547/edcdb1e5-ed9d-49e6-97e6-ea7629682547.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1020.957728] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1020.958699] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d6ce510-ae2e-4e0a-b292-fc1bd335e262 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.967028] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1020.967028] env[62814]: value = "task-4294231" [ 1020.967028] env[62814]: _type = "Task" [ 1020.967028] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.982433] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b636e2-ac8d-9abe-5fc3-87b332f6c46c, 'name': SearchDatastore_Task, 'duration_secs': 0.012458} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.986744] env[62814]: DEBUG oslo_concurrency.lockutils [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1020.987087] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] e727eeb1-c5d0-4591-80bb-31746bf976a2/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. {{(pid=62814) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1020.987416] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294231, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.987660] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b001a8f4-f9a9-4ddf-97b1-0aeed6de6178 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.991317] env[62814]: DEBUG nova.policy [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e246f49b0b84cd093549b6d6b384e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6e52480dd2c467790622901940cf385', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1020.998517] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1020.998517] env[62814]: value = "task-4294232" [ 1020.998517] env[62814]: _type = "Task" [ 1020.998517] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.009613] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294232, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.152400] env[62814]: DEBUG oslo_concurrency.lockutils [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "508321ab-ce10-4953-a9e3-193b9975bec7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1021.154255] env[62814]: DEBUG oslo_concurrency.lockutils [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "508321ab-ce10-4953-a9e3-193b9975bec7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1021.154567] env[62814]: DEBUG oslo_concurrency.lockutils [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "508321ab-ce10-4953-a9e3-193b9975bec7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1021.154737] env[62814]: DEBUG oslo_concurrency.lockutils [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "508321ab-ce10-4953-a9e3-193b9975bec7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1021.154917] env[62814]: DEBUG oslo_concurrency.lockutils [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "508321ab-ce10-4953-a9e3-193b9975bec7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1021.157259] env[62814]: INFO nova.compute.manager [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Terminating instance [ 1021.413087] env[62814]: DEBUG nova.compute.manager [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1021.451888] env[62814]: DEBUG nova.network.neutron [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Successfully created port: 3f437142-559d-418d-814f-10278dd5ed06 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1021.485471] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294231, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075492} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.488679] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1021.489963] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02792044-635f-45c4-a9a5-0a8e1b213d3d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.525364] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] edcdb1e5-ed9d-49e6-97e6-ea7629682547/edcdb1e5-ed9d-49e6-97e6-ea7629682547.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.533566] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-333be767-8eee-4125-b839-a8bdfae0dfd2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.559491] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294232, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533542} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.560996] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] e727eeb1-c5d0-4591-80bb-31746bf976a2/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. [ 1021.561671] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1021.561671] env[62814]: value = "task-4294234" [ 1021.561671] env[62814]: _type = "Task" [ 1021.561671] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.562460] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37e3a99-7b37-4049-93f6-9f7bdda0dfde {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.585062] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294234, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.623825] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] e727eeb1-c5d0-4591-80bb-31746bf976a2/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.624755] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7b4ba00-d63f-40cb-bb56-5927969090b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.661344] env[62814]: DEBUG nova.compute.manager [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1021.661783] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1021.664358] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e6e359-4f4b-42c5-a597-caaee1ea21bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.669849] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1021.669849] env[62814]: value = "task-4294235" [ 1021.669849] env[62814]: _type = "Task" [ 1021.669849] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.682199] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.683297] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c638310e-4d7a-446b-9ca8-324320730bd2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.688723] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.698080] env[62814]: DEBUG oslo_vmware.api [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 1021.698080] env[62814]: value = "task-4294236" [ 1021.698080] env[62814]: _type = "Task" [ 1021.698080] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.708316] env[62814]: DEBUG oslo_vmware.api [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294236, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.051022] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c0543f-d154-42e7-8da9-c2c95abfcdd5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.064056] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f84240-2d40-4c75-be05-497bfe0b2ed6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.102583] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e250f974-1764-4df0-8085-63bbcc73f4ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.110342] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294234, 'name': ReconfigVM_Task, 'duration_secs': 0.338363} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.111294] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Reconfigured VM instance instance-00000050 to attach disk [datastore2] edcdb1e5-ed9d-49e6-97e6-ea7629682547/edcdb1e5-ed9d-49e6-97e6-ea7629682547.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.112644] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f7eaa3eb-f23d-4bc4-86f9-77b945bf2833 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.117278] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ac1101-916c-4f35-9055-c02aeb26ac97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.122809] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1022.122809] env[62814]: value = "task-4294237" [ 1022.122809] env[62814]: _type = "Task" [ 1022.122809] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.135792] env[62814]: DEBUG nova.compute.provider_tree [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.143354] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294237, 'name': Rename_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.184925] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294235, 'name': ReconfigVM_Task, 'duration_secs': 0.348416} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.185283] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Reconfigured VM instance instance-0000004f to attach disk [datastore2] e727eeb1-c5d0-4591-80bb-31746bf976a2/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.186262] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0658a0c-f188-48f4-8d1f-3f79b5c0924b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.217621] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5eda6c82-f2dd-448e-a3aa-50e126c2b1b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.238235] env[62814]: DEBUG oslo_vmware.api [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294236, 'name': PowerOffVM_Task, 'duration_secs': 0.200388} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.239891] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.240050] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1022.240305] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1022.240305] env[62814]: value = "task-4294238" [ 1022.240305] env[62814]: _type = "Task" [ 1022.240305] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.240513] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-525d1ea4-1edc-4c92-9213-3ac4af973661 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.254769] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294238, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.318396] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1022.318512] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1022.318701] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Deleting the datastore file [datastore2] 508321ab-ce10-4953-a9e3-193b9975bec7 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.318982] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f2a9188-702f-4f62-b8e2-65ff103cd59b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.329499] env[62814]: DEBUG oslo_vmware.api [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 1022.329499] env[62814]: value = "task-4294240" [ 1022.329499] env[62814]: _type = "Task" [ 1022.329499] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.340924] env[62814]: DEBUG oslo_vmware.api [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294240, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.433831] env[62814]: DEBUG nova.compute.manager [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1022.475062] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1022.475326] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.475469] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1022.475659] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.475802] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1022.475947] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1022.476212] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1022.476398] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1022.476596] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1022.476779] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1022.476987] env[62814]: DEBUG nova.virt.hardware [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1022.477964] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a40ff67-f947-4752-8a09-fc19eb1cfde1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.487991] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33dbb7a0-5dc0-4b31-8b0f-68b047bfea14 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.633910] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294237, 'name': Rename_Task, 'duration_secs': 0.294749} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.634237] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.634486] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42fc003e-3f45-43f6-b8fe-7bc5ebd49d5f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.639376] env[62814]: DEBUG nova.scheduler.client.report [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1022.644393] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1022.644393] env[62814]: value = "task-4294241" [ 1022.644393] env[62814]: _type = "Task" [ 1022.644393] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.658911] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294241, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.754430] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294238, 'name': ReconfigVM_Task, 'duration_secs': 0.351434} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.754588] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.754844] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-734beb96-629d-45f4-8ce8-30cfd7e6bf85 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.762954] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1022.762954] env[62814]: value = "task-4294242" [ 1022.762954] env[62814]: _type = "Task" [ 1022.762954] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.774795] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294242, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.841338] env[62814]: DEBUG oslo_vmware.api [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294240, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.392477} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.841587] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.841817] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1022.841984] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1022.842179] env[62814]: INFO nova.compute.manager [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1022.842451] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1022.842665] env[62814]: DEBUG nova.compute.manager [-] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1022.842764] env[62814]: DEBUG nova.network.neutron [-] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1023.151676] env[62814]: DEBUG nova.compute.manager [req-4049d878-313f-4b34-9706-2b3e20453c24 req-8f8ed2ec-894f-42e4-b76d-79876efeac9a service nova] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Received event network-vif-plugged-3f437142-559d-418d-814f-10278dd5ed06 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1023.151937] env[62814]: DEBUG oslo_concurrency.lockutils [req-4049d878-313f-4b34-9706-2b3e20453c24 req-8f8ed2ec-894f-42e4-b76d-79876efeac9a service nova] Acquiring lock "316931ae-7a62-4bac-81e4-1fee9a36164e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1023.152304] env[62814]: DEBUG oslo_concurrency.lockutils [req-4049d878-313f-4b34-9706-2b3e20453c24 req-8f8ed2ec-894f-42e4-b76d-79876efeac9a service nova] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.153318] env[62814]: DEBUG oslo_concurrency.lockutils [req-4049d878-313f-4b34-9706-2b3e20453c24 req-8f8ed2ec-894f-42e4-b76d-79876efeac9a service nova] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.153318] env[62814]: DEBUG nova.compute.manager [req-4049d878-313f-4b34-9706-2b3e20453c24 req-8f8ed2ec-894f-42e4-b76d-79876efeac9a service nova] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] No waiting events found dispatching network-vif-plugged-3f437142-559d-418d-814f-10278dd5ed06 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1023.153394] env[62814]: WARNING nova.compute.manager [req-4049d878-313f-4b34-9706-2b3e20453c24 req-8f8ed2ec-894f-42e4-b76d-79876efeac9a service nova] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Received unexpected event network-vif-plugged-3f437142-559d-418d-814f-10278dd5ed06 for instance with vm_state building and task_state spawning. [ 1023.154182] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.755s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1023.154649] env[62814]: DEBUG nova.compute.manager [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1023.160980] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 20.188s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1023.173868] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294241, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.259592] env[62814]: DEBUG nova.network.neutron [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Successfully updated port: 3f437142-559d-418d-814f-10278dd5ed06 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1023.276700] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294242, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.296598] env[62814]: DEBUG nova.compute.manager [req-36d3aa80-ccf5-4c69-a5ab-08f0448d2d4b req-4cb6924e-1b4c-44af-9f8b-714eacd09eec service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Received event network-vif-deleted-5992ceda-be00-450a-a26e-10cab2e2a02b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1023.296773] env[62814]: INFO nova.compute.manager [req-36d3aa80-ccf5-4c69-a5ab-08f0448d2d4b req-4cb6924e-1b4c-44af-9f8b-714eacd09eec service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Neutron deleted interface 5992ceda-be00-450a-a26e-10cab2e2a02b; detaching it from the instance and deleting it from the info cache [ 1023.297016] env[62814]: DEBUG nova.network.neutron [req-36d3aa80-ccf5-4c69-a5ab-08f0448d2d4b req-4cb6924e-1b4c-44af-9f8b-714eacd09eec service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.660474] env[62814]: DEBUG oslo_vmware.api [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294241, 'name': PowerOnVM_Task, 'duration_secs': 0.703832} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.660710] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1023.660957] env[62814]: INFO nova.compute.manager [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Took 8.18 seconds to spawn the instance on the hypervisor. [ 1023.661164] env[62814]: DEBUG nova.compute.manager [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1023.662031] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de7974f-b6af-4875-842a-caa1d7569560 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.665577] env[62814]: DEBUG nova.compute.utils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1023.668872] env[62814]: INFO nova.compute.claims [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.672507] env[62814]: DEBUG nova.compute.manager [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1023.672681] env[62814]: DEBUG nova.network.neutron [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1023.726027] env[62814]: DEBUG nova.policy [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd71116c0cfd4b509e4070ecc84613f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd82993ef7dfa4d2f8f39db3577dd321a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1023.731009] env[62814]: DEBUG nova.network.neutron [-] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.765687] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "refresh_cache-316931ae-7a62-4bac-81e4-1fee9a36164e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.765851] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "refresh_cache-316931ae-7a62-4bac-81e4-1fee9a36164e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1023.766023] env[62814]: DEBUG nova.network.neutron [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1023.777536] env[62814]: DEBUG oslo_vmware.api [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294242, 'name': PowerOnVM_Task, 'duration_secs': 0.523546} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.779133] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1023.781882] env[62814]: DEBUG nova.compute.manager [None req-db41a72e-d2ed-4e3c-8deb-7a23d0998d50 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1023.782849] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9b5706-4290-4d38-8d44-ca7ca061bfa9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.802054] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df15b60e-1550-49f9-94a7-43c41d054615 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.813154] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b19c501-be0c-46bd-b8a7-8e6bdd30ee89 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.850137] env[62814]: DEBUG nova.compute.manager [req-36d3aa80-ccf5-4c69-a5ab-08f0448d2d4b req-4cb6924e-1b4c-44af-9f8b-714eacd09eec service nova] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Detach interface failed, port_id=5992ceda-be00-450a-a26e-10cab2e2a02b, reason: Instance 508321ab-ce10-4953-a9e3-193b9975bec7 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1024.073526] env[62814]: DEBUG nova.network.neutron [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Successfully created port: af952a49-d433-42b8-8a05-fc3d6e780333 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1024.173092] env[62814]: DEBUG nova.compute.manager [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1024.177390] env[62814]: INFO nova.compute.resource_tracker [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating resource usage from migration 06826230-b161-48d6-bc89-eabd48b17ffe [ 1024.197749] env[62814]: INFO nova.compute.manager [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Took 34.00 seconds to build instance. [ 1024.238674] env[62814]: INFO nova.compute.manager [-] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Took 1.39 seconds to deallocate network for instance. [ 1024.308936] env[62814]: DEBUG nova.network.neutron [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1024.506632] env[62814]: DEBUG nova.network.neutron [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Updating instance_info_cache with network_info: [{"id": "3f437142-559d-418d-814f-10278dd5ed06", "address": "fa:16:3e:69:59:4f", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f437142-55", "ovs_interfaceid": "3f437142-559d-418d-814f-10278dd5ed06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.610047] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63960d17-cc18-43b8-ad82-f252e1b113ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.623913] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-118f43fd-3d00-4975-9b79-65b3b0c5c882 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.657884] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0079509-3ea3-487c-868a-39575d830264 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.667232] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-780b9cb6-24b9-43dd-947b-7dbdec0208c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.684066] env[62814]: DEBUG nova.compute.provider_tree [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.703406] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6d3e0c20-a829-4015-85f7-caf56a069f79 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.528s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1024.744615] env[62814]: DEBUG oslo_concurrency.lockutils [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1025.011719] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "refresh_cache-316931ae-7a62-4bac-81e4-1fee9a36164e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1025.012079] env[62814]: DEBUG nova.compute.manager [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Instance network_info: |[{"id": "3f437142-559d-418d-814f-10278dd5ed06", "address": "fa:16:3e:69:59:4f", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f437142-55", "ovs_interfaceid": "3f437142-559d-418d-814f-10278dd5ed06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1025.012513] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:59:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '51bac3c3-00ab-4a07-9e28-b3c951dee565', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f437142-559d-418d-814f-10278dd5ed06', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.020215] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1025.020475] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1025.020728] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-445890a1-6a3e-4832-8cbc-c6c4a425126e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.045507] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.045507] env[62814]: value = "task-4294244" [ 1025.045507] env[62814]: _type = "Task" [ 1025.045507] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.056745] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294244, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.178745] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "cd037f6e-fddd-4389-b6d0-144b798537bf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1025.179025] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "cd037f6e-fddd-4389-b6d0-144b798537bf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.189334] env[62814]: DEBUG nova.compute.manager [req-07439f95-4a91-4c0e-9549-a1f66091ddc4 req-ff5ae70e-c949-4bdd-aa19-ad24593f03b7 service nova] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Received event network-changed-3f437142-559d-418d-814f-10278dd5ed06 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1025.189334] env[62814]: DEBUG nova.compute.manager [req-07439f95-4a91-4c0e-9549-a1f66091ddc4 req-ff5ae70e-c949-4bdd-aa19-ad24593f03b7 service nova] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Refreshing instance network info cache due to event network-changed-3f437142-559d-418d-814f-10278dd5ed06. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1025.189334] env[62814]: DEBUG oslo_concurrency.lockutils [req-07439f95-4a91-4c0e-9549-a1f66091ddc4 req-ff5ae70e-c949-4bdd-aa19-ad24593f03b7 service nova] Acquiring lock "refresh_cache-316931ae-7a62-4bac-81e4-1fee9a36164e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.189334] env[62814]: DEBUG oslo_concurrency.lockutils [req-07439f95-4a91-4c0e-9549-a1f66091ddc4 req-ff5ae70e-c949-4bdd-aa19-ad24593f03b7 service nova] Acquired lock "refresh_cache-316931ae-7a62-4bac-81e4-1fee9a36164e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1025.189334] env[62814]: DEBUG nova.network.neutron [req-07439f95-4a91-4c0e-9549-a1f66091ddc4 req-ff5ae70e-c949-4bdd-aa19-ad24593f03b7 service nova] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Refreshing network info cache for port 3f437142-559d-418d-814f-10278dd5ed06 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1025.191039] env[62814]: DEBUG nova.scheduler.client.report [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1025.195631] env[62814]: DEBUG nova.compute.manager [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1025.234881] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1025.235289] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.235478] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1025.235673] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.235819] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1025.235964] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1025.236193] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1025.236348] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1025.236519] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1025.236686] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1025.236865] env[62814]: DEBUG nova.virt.hardware [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1025.237760] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be369dce-536a-4756-8b98-8f226561521e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.247564] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd919a1-9e36-47a9-9b84-43234ea2cfdd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.561947] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294244, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.681773] env[62814]: DEBUG nova.compute.manager [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1025.699234] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.538s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1025.699485] env[62814]: INFO nova.compute.manager [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Migrating [ 1025.711129] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.031s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1025.712900] env[62814]: INFO nova.compute.claims [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.772695] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.773054] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.773054] env[62814]: DEBUG nova.network.neutron [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1026.775104] env[62814]: DEBUG nova.network.neutron [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Successfully updated port: af952a49-d433-42b8-8a05-fc3d6e780333 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1026.781190] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.781190] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1026.797042] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294244, 'name': CreateVM_Task, 'duration_secs': 1.378235} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.797042] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1026.797042] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.797042] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1026.797042] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1026.797042] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-562b56b4-1dbd-4514-a2e4-1288b2199592 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.801906] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1026.801906] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc9837-247d-4913-ec9a-d5c8f56962d1" [ 1026.801906] env[62814]: _type = "Task" [ 1026.801906] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.807163] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1026.814210] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc9837-247d-4913-ec9a-d5c8f56962d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.860909] env[62814]: DEBUG nova.network.neutron [req-07439f95-4a91-4c0e-9549-a1f66091ddc4 req-ff5ae70e-c949-4bdd-aa19-ad24593f03b7 service nova] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Updated VIF entry in instance network info cache for port 3f437142-559d-418d-814f-10278dd5ed06. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1026.861092] env[62814]: DEBUG nova.network.neutron [req-07439f95-4a91-4c0e-9549-a1f66091ddc4 req-ff5ae70e-c949-4bdd-aa19-ad24593f03b7 service nova] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Updating instance_info_cache with network_info: [{"id": "3f437142-559d-418d-814f-10278dd5ed06", "address": "fa:16:3e:69:59:4f", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f437142-55", "ovs_interfaceid": "3f437142-559d-418d-814f-10278dd5ed06", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.174851] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Acquiring lock "d5df1d0e-55e6-452e-882a-053a83250b39" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.175126] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "d5df1d0e-55e6-452e-882a-053a83250b39" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.289101] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "refresh_cache-4ebc3280-64a7-4e9b-8385-2128357d3422" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.289101] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "refresh_cache-4ebc3280-64a7-4e9b-8385-2128357d3422" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1027.289101] env[62814]: DEBUG nova.network.neutron [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.293370] env[62814]: DEBUG nova.compute.manager [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1027.297949] env[62814]: DEBUG nova.compute.manager [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Received event network-vif-plugged-af952a49-d433-42b8-8a05-fc3d6e780333 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1027.298248] env[62814]: DEBUG oslo_concurrency.lockutils [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] Acquiring lock "4ebc3280-64a7-4e9b-8385-2128357d3422-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.298479] env[62814]: DEBUG oslo_concurrency.lockutils [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] Lock "4ebc3280-64a7-4e9b-8385-2128357d3422-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.298650] env[62814]: DEBUG oslo_concurrency.lockutils [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] Lock "4ebc3280-64a7-4e9b-8385-2128357d3422-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1027.298903] env[62814]: DEBUG nova.compute.manager [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] No waiting events found dispatching network-vif-plugged-af952a49-d433-42b8-8a05-fc3d6e780333 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1027.298971] env[62814]: WARNING nova.compute.manager [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Received unexpected event network-vif-plugged-af952a49-d433-42b8-8a05-fc3d6e780333 for instance with vm_state building and task_state spawning. [ 1027.299142] env[62814]: DEBUG nova.compute.manager [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Received event network-changed-af952a49-d433-42b8-8a05-fc3d6e780333 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1027.299293] env[62814]: DEBUG nova.compute.manager [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Refreshing instance network info cache due to event network-changed-af952a49-d433-42b8-8a05-fc3d6e780333. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1027.299455] env[62814]: DEBUG oslo_concurrency.lockutils [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] Acquiring lock "refresh_cache-4ebc3280-64a7-4e9b-8385-2128357d3422" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.316145] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc9837-247d-4913-ec9a-d5c8f56962d1, 'name': SearchDatastore_Task, 'duration_secs': 0.013916} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.316934] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1027.317155] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.317391] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.317535] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1027.317713] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.317982] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c9086b0-6708-4d08-b807-9dcfa7144d90 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.336855] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1027.337323] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1027.337820] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c06f178-3580-447b-8a3b-021640199669 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.349512] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1027.349512] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5208cef4-3391-3a6f-cd6f-8b1570937e27" [ 1027.349512] env[62814]: _type = "Task" [ 1027.349512] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.359181] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5208cef4-3391-3a6f-cd6f-8b1570937e27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.364654] env[62814]: DEBUG oslo_concurrency.lockutils [req-07439f95-4a91-4c0e-9549-a1f66091ddc4 req-ff5ae70e-c949-4bdd-aa19-ad24593f03b7 service nova] Releasing lock "refresh_cache-316931ae-7a62-4bac-81e4-1fee9a36164e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1027.618305] env[62814]: DEBUG nova.network.neutron [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance_info_cache with network_info: [{"id": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "address": "fa:16:3e:db:ee:dc", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08bab1d2-62", "ovs_interfaceid": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.714085] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquiring lock "44ea319c-6ea0-456a-bee6-42133a25d8c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.714085] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lock "44ea319c-6ea0-456a-bee6-42133a25d8c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1027.770119] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f13004b-9760-43c3-8c8e-ed6897c152fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.778911] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70cb1bd5-fd9e-4954-a94e-985173db0e77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.813453] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a81364a-37ea-4697-a778-2629829b40f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.824204] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67b78261-d8b3-4e34-879a-ba5142d2d949 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.839019] env[62814]: DEBUG nova.compute.provider_tree [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.841058] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1027.859196] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5208cef4-3391-3a6f-cd6f-8b1570937e27, 'name': SearchDatastore_Task, 'duration_secs': 0.049754} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.859962] env[62814]: DEBUG nova.network.neutron [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1027.862427] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06e77db0-64ce-4954-8642-c57e5987ba37 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.868594] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1027.868594] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b67f94-2ade-99b2-51d4-9e7201e0e2d4" [ 1027.868594] env[62814]: _type = "Task" [ 1027.868594] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.877601] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b67f94-2ade-99b2-51d4-9e7201e0e2d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.996522] env[62814]: DEBUG nova.network.neutron [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Updating instance_info_cache with network_info: [{"id": "af952a49-d433-42b8-8a05-fc3d6e780333", "address": "fa:16:3e:02:3f:68", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf952a49-d4", "ovs_interfaceid": "af952a49-d433-42b8-8a05-fc3d6e780333", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.119169] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.343542] env[62814]: DEBUG nova.scheduler.client.report [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1028.379740] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b67f94-2ade-99b2-51d4-9e7201e0e2d4, 'name': SearchDatastore_Task, 'duration_secs': 0.028169} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.380016] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.380300] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 316931ae-7a62-4bac-81e4-1fee9a36164e/316931ae-7a62-4bac-81e4-1fee9a36164e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1028.380627] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-141cb78d-9442-4d85-82e2-5140121c2c19 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.389399] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1028.389399] env[62814]: value = "task-4294245" [ 1028.389399] env[62814]: _type = "Task" [ 1028.389399] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.398674] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294245, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.500027] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "refresh_cache-4ebc3280-64a7-4e9b-8385-2128357d3422" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1028.500027] env[62814]: DEBUG nova.compute.manager [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Instance network_info: |[{"id": "af952a49-d433-42b8-8a05-fc3d6e780333", "address": "fa:16:3e:02:3f:68", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf952a49-d4", "ovs_interfaceid": "af952a49-d433-42b8-8a05-fc3d6e780333", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1028.500027] env[62814]: DEBUG oslo_concurrency.lockutils [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] Acquired lock "refresh_cache-4ebc3280-64a7-4e9b-8385-2128357d3422" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1028.500367] env[62814]: DEBUG nova.network.neutron [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Refreshing network info cache for port af952a49-d433-42b8-8a05-fc3d6e780333 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1028.505021] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:3f:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'af952a49-d433-42b8-8a05-fc3d6e780333', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1028.509245] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1028.510270] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1028.510507] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-56241fb1-0d7f-4506-b187-3415350e4034 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.536456] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1028.536456] env[62814]: value = "task-4294246" [ 1028.536456] env[62814]: _type = "Task" [ 1028.536456] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.545177] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294246, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.849182] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.138s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1028.849832] env[62814]: DEBUG nova.compute.manager [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1028.853723] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.426s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1028.853961] env[62814]: DEBUG nova.objects.instance [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lazy-loading 'resources' on Instance uuid b5ddd05f-2027-4edf-84c5-0d2f537a95e4 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.903214] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294245, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.048720] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294246, 'name': CreateVM_Task, 'duration_secs': 0.426699} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.048870] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1029.049552] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.049719] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.050047] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1029.050300] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b23c031-b7d6-46a4-800f-b37b0cad160b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.055238] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1029.055238] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f750ac-5855-4ba2-bfe8-4105a145f945" [ 1029.055238] env[62814]: _type = "Task" [ 1029.055238] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.066501] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f750ac-5855-4ba2-bfe8-4105a145f945, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.245280] env[62814]: DEBUG nova.network.neutron [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Updated VIF entry in instance network info cache for port af952a49-d433-42b8-8a05-fc3d6e780333. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1029.245662] env[62814]: DEBUG nova.network.neutron [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Updating instance_info_cache with network_info: [{"id": "af952a49-d433-42b8-8a05-fc3d6e780333", "address": "fa:16:3e:02:3f:68", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf952a49-d4", "ovs_interfaceid": "af952a49-d433-42b8-8a05-fc3d6e780333", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.358259] env[62814]: DEBUG nova.compute.utils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1029.365912] env[62814]: DEBUG nova.compute.manager [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1029.366343] env[62814]: DEBUG nova.network.neutron [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1029.403232] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294245, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679096} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.403511] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 316931ae-7a62-4bac-81e4-1fee9a36164e/316931ae-7a62-4bac-81e4-1fee9a36164e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1029.403736] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.403985] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aad10f2f-abf3-4530-baf1-46f4fa89f9a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.411937] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1029.411937] env[62814]: value = "task-4294247" [ 1029.411937] env[62814]: _type = "Task" [ 1029.411937] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.415526] env[62814]: DEBUG nova.policy [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64e84da705284d2dbf693a26ef184cd7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95c92336f9e746edba50b0b9e078b0dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1029.423917] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294247, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.569058] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f750ac-5855-4ba2-bfe8-4105a145f945, 'name': SearchDatastore_Task, 'duration_secs': 0.02006} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.569385] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1029.569654] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1029.569894] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.570059] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1029.570235] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1029.570497] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e011ec02-987e-44b2-9118-7c38ed549a4c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.581144] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1029.581787] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1029.584859] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c3dc64d-efc0-4ef3-b57d-58679bce1cac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.591346] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1029.591346] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5205cf75-0313-0309-d969-37fe516aee5a" [ 1029.591346] env[62814]: _type = "Task" [ 1029.591346] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.600013] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5205cf75-0313-0309-d969-37fe516aee5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.638105] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2737983-f7b2-4046-980f-361c83f9273e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.657936] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance '9343301c-dfe9-41b0-b4a0-067af544d297' progress to 0 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1029.752595] env[62814]: DEBUG oslo_concurrency.lockutils [req-b94da3dc-004d-4933-aba8-b1e40431568f req-26da34ac-3e37-4677-9de9-a0e118512f2e service nova] Releasing lock "refresh_cache-4ebc3280-64a7-4e9b-8385-2128357d3422" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1029.761777] env[62814]: DEBUG nova.network.neutron [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Successfully created port: 08e04ba7-4665-4982-ac59-6412cc953d93 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1029.916746] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f46047-eff4-43db-ab73-3c9d334d8d8c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.916746] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e898539-28e5-40bf-a0ca-0d70074ec3fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.916746] env[62814]: DEBUG nova.compute.manager [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1029.916746] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a56594c-30ac-4d3a-89bc-fb5050a6c304 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.916746] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e0a472b-2e4d-47b8-a6e4-44b112a6acf9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.919465] env[62814]: DEBUG nova.compute.provider_tree [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.932882] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294247, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.204468} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.934071] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1029.934689] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658941dd-454f-4611-a341-86e4a553e974 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.958647] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 316931ae-7a62-4bac-81e4-1fee9a36164e/316931ae-7a62-4bac-81e4-1fee9a36164e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1029.959162] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2f7ab178-c4da-4542-9541-b836526a5a58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.980171] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1029.980171] env[62814]: value = "task-4294248" [ 1029.980171] env[62814]: _type = "Task" [ 1029.980171] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.988884] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294248, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.102992] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5205cf75-0313-0309-d969-37fe516aee5a, 'name': SearchDatastore_Task, 'duration_secs': 0.014325} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.103857] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba04685d-04fe-4cac-a21b-92996c2c06f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.110777] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1030.110777] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c71606-ce9b-5f3e-f8d7-3a51e5345922" [ 1030.110777] env[62814]: _type = "Task" [ 1030.110777] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.120044] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c71606-ce9b-5f3e-f8d7-3a51e5345922, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.166103] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1030.166469] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-307d7c0c-8698-4ebc-a60c-a69d9431ded6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.176162] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1030.176162] env[62814]: value = "task-4294249" [ 1030.176162] env[62814]: _type = "Task" [ 1030.176162] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.185564] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294249, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.422542] env[62814]: DEBUG nova.scheduler.client.report [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1030.490284] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294248, 'name': ReconfigVM_Task, 'duration_secs': 0.291637} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.490570] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 316931ae-7a62-4bac-81e4-1fee9a36164e/316931ae-7a62-4bac-81e4-1fee9a36164e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.491212] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-132727cf-c9d2-45f4-a0e8-4df9599aae6b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.498801] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1030.498801] env[62814]: value = "task-4294250" [ 1030.498801] env[62814]: _type = "Task" [ 1030.498801] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.507728] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294250, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.623675] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c71606-ce9b-5f3e-f8d7-3a51e5345922, 'name': SearchDatastore_Task, 'duration_secs': 0.012334} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.624100] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1030.624419] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 4ebc3280-64a7-4e9b-8385-2128357d3422/4ebc3280-64a7-4e9b-8385-2128357d3422.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1030.624744] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d70c4ad4-72c2-4a45-a9f1-87419eba44dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.633740] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1030.633740] env[62814]: value = "task-4294251" [ 1030.633740] env[62814]: _type = "Task" [ 1030.633740] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.645225] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294251, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.687564] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1030.687848] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance '9343301c-dfe9-41b0-b4a0-067af544d297' progress to 17 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1030.727057] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "9683894b-a300-4400-a1b9-db62478f42c5" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1030.727397] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1030.897194] env[62814]: DEBUG nova.compute.manager [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1030.925623] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1030.925924] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.926109] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1030.926367] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.926519] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1030.927527] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1030.927527] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1030.927527] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1030.927527] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1030.927527] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1030.927812] env[62814]: DEBUG nova.virt.hardware [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1030.928676] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.075s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1030.931930] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5685370-60de-409a-8843-f681f1aa4243 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.935547] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.557s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1030.937511] env[62814]: INFO nova.compute.claims [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1030.948343] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc65477-1032-44e9-ade3-482974089ba2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.966977] env[62814]: INFO nova.scheduler.client.report [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted allocations for instance b5ddd05f-2027-4edf-84c5-0d2f537a95e4 [ 1031.011885] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294250, 'name': Rename_Task, 'duration_secs': 0.150629} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.012613] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1031.012613] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f56d4524-5a31-4191-96ce-c9e302ef6cbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.023070] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1031.023070] env[62814]: value = "task-4294252" [ 1031.023070] env[62814]: _type = "Task" [ 1031.023070] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.038162] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294252, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.147822] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294251, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.194972] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1031.195267] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.195407] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1031.195589] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.195737] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1031.195882] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1031.196104] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1031.196264] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1031.196430] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1031.196591] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1031.196758] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1031.202015] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10555768-2fdf-4077-a008-06d6acf16712 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.221744] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1031.221744] env[62814]: value = "task-4294253" [ 1031.221744] env[62814]: _type = "Task" [ 1031.221744] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.234285] env[62814]: INFO nova.compute.manager [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Detaching volume f4d241a2-9ed8-46ae-af01-e7e67273236c [ 1031.236342] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294253, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.293849] env[62814]: INFO nova.virt.block_device [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Attempting to driver detach volume f4d241a2-9ed8-46ae-af01-e7e67273236c from mountpoint /dev/sdb [ 1031.294147] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1031.294328] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845759', 'volume_id': 'f4d241a2-9ed8-46ae-af01-e7e67273236c', 'name': 'volume-f4d241a2-9ed8-46ae-af01-e7e67273236c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9683894b-a300-4400-a1b9-db62478f42c5', 'attached_at': '', 'detached_at': '', 'volume_id': 'f4d241a2-9ed8-46ae-af01-e7e67273236c', 'serial': 'f4d241a2-9ed8-46ae-af01-e7e67273236c'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1031.295230] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3189a3b-3e48-428b-8a5a-52c585585707 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.328647] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80cbbfd-6901-4c23-93fb-fcf03f280067 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.338781] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c3886c-523f-4d4c-8552-b89fd51a112b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.378822] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6926c900-3006-4f12-ab2e-eddbca8480d6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.396538] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] The volume has not been displaced from its original location: [datastore1] volume-f4d241a2-9ed8-46ae-af01-e7e67273236c/volume-f4d241a2-9ed8-46ae-af01-e7e67273236c.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1031.402802] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Reconfiguring VM instance instance-00000035 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1031.403029] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-520d340d-34ba-4641-8603-482c6ac21d41 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.422009] env[62814]: DEBUG nova.compute.manager [req-889ba1b3-b28a-46b3-8b31-ad6d92129ec6 req-f76ec7f6-4f2b-4c2b-a7dc-8351b5be80a9 service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Received event network-vif-plugged-08e04ba7-4665-4982-ac59-6412cc953d93 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1031.422329] env[62814]: DEBUG oslo_concurrency.lockutils [req-889ba1b3-b28a-46b3-8b31-ad6d92129ec6 req-f76ec7f6-4f2b-4c2b-a7dc-8351b5be80a9 service nova] Acquiring lock "a07b964f-fe88-4c88-b6f4-61ed0973716c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1031.422528] env[62814]: DEBUG oslo_concurrency.lockutils [req-889ba1b3-b28a-46b3-8b31-ad6d92129ec6 req-f76ec7f6-4f2b-4c2b-a7dc-8351b5be80a9 service nova] Lock "a07b964f-fe88-4c88-b6f4-61ed0973716c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1031.422632] env[62814]: DEBUG oslo_concurrency.lockutils [req-889ba1b3-b28a-46b3-8b31-ad6d92129ec6 req-f76ec7f6-4f2b-4c2b-a7dc-8351b5be80a9 service nova] Lock "a07b964f-fe88-4c88-b6f4-61ed0973716c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.422794] env[62814]: DEBUG nova.compute.manager [req-889ba1b3-b28a-46b3-8b31-ad6d92129ec6 req-f76ec7f6-4f2b-4c2b-a7dc-8351b5be80a9 service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] No waiting events found dispatching network-vif-plugged-08e04ba7-4665-4982-ac59-6412cc953d93 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1031.422951] env[62814]: WARNING nova.compute.manager [req-889ba1b3-b28a-46b3-8b31-ad6d92129ec6 req-f76ec7f6-4f2b-4c2b-a7dc-8351b5be80a9 service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Received unexpected event network-vif-plugged-08e04ba7-4665-4982-ac59-6412cc953d93 for instance with vm_state building and task_state spawning. [ 1031.424953] env[62814]: DEBUG oslo_vmware.api [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1031.424953] env[62814]: value = "task-4294254" [ 1031.424953] env[62814]: _type = "Task" [ 1031.424953] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.436631] env[62814]: DEBUG oslo_vmware.api [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294254, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.475554] env[62814]: DEBUG oslo_concurrency.lockutils [None req-81e52a1d-8d3f-447d-b707-d6118129230c tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "b5ddd05f-2027-4edf-84c5-0d2f537a95e4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.744s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1031.514186] env[62814]: DEBUG nova.network.neutron [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Successfully updated port: 08e04ba7-4665-4982-ac59-6412cc953d93 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.535829] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294252, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.649106] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294251, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.74541} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.649863] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 4ebc3280-64a7-4e9b-8385-2128357d3422/4ebc3280-64a7-4e9b-8385-2128357d3422.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1031.649863] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1031.649863] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df66a933-dc52-4bfb-adcd-424b0ccdd173 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.659503] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1031.659503] env[62814]: value = "task-4294255" [ 1031.659503] env[62814]: _type = "Task" [ 1031.659503] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.670219] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294255, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.732496] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294253, 'name': ReconfigVM_Task, 'duration_secs': 0.279144} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.732865] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance '9343301c-dfe9-41b0-b4a0-067af544d297' progress to 33 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1031.940397] env[62814]: DEBUG oslo_vmware.api [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294254, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.018049] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-a07b964f-fe88-4c88-b6f4-61ed0973716c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.018049] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-a07b964f-fe88-4c88-b6f4-61ed0973716c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1032.018049] env[62814]: DEBUG nova.network.neutron [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1032.038406] env[62814]: DEBUG oslo_vmware.api [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294252, 'name': PowerOnVM_Task, 'duration_secs': 0.741885} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.042141] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1032.042141] env[62814]: INFO nova.compute.manager [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Took 9.61 seconds to spawn the instance on the hypervisor. [ 1032.042282] env[62814]: DEBUG nova.compute.manager [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1032.044319] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51212d8b-15c1-43d4-9ad2-4bf3ef5a4cad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.169352] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294255, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.105281} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.171932] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1032.173423] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d81376-bf39-4719-a803-c1286fbb5979 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.199711] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 4ebc3280-64a7-4e9b-8385-2128357d3422/4ebc3280-64a7-4e9b-8385-2128357d3422.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.201871] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0a64012-ac74-43d8-a908-9db25eb21040 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.223718] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1032.223718] env[62814]: value = "task-4294256" [ 1032.223718] env[62814]: _type = "Task" [ 1032.223718] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.236473] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294256, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.241654] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1032.241928] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1032.242098] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1032.242283] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1032.242446] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1032.242848] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1032.242848] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1032.242956] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1032.243125] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1032.243286] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1032.243456] env[62814]: DEBUG nova.virt.hardware [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1032.248746] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Reconfiguring VM instance instance-00000023 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1032.251530] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e890a44d-d980-493b-9719-89b564fdbe2c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.271821] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1032.271821] env[62814]: value = "task-4294257" [ 1032.271821] env[62814]: _type = "Task" [ 1032.271821] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.283985] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294257, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.395373] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963ac890-8bae-4f8e-a76b-8a8365f895be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.405357] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab79df08-1712-4600-a410-c26040858719 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.438784] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f026a4-ab82-44c9-8125-9c6af94e9aa2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.449758] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fca2f378-55af-4aca-b78e-9e1b6146fd00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.453455] env[62814]: DEBUG oslo_vmware.api [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294254, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.463956] env[62814]: DEBUG nova.compute.provider_tree [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.564421] env[62814]: DEBUG nova.network.neutron [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.568536] env[62814]: INFO nova.compute.manager [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Took 31.76 seconds to build instance. [ 1032.737946] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294256, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.771267] env[62814]: DEBUG nova.network.neutron [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Updating instance_info_cache with network_info: [{"id": "08e04ba7-4665-4982-ac59-6412cc953d93", "address": "fa:16:3e:a3:d9:48", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08e04ba7-46", "ovs_interfaceid": "08e04ba7-4665-4982-ac59-6412cc953d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.782320] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294257, 'name': ReconfigVM_Task, 'duration_secs': 0.400635} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.782593] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Reconfigured VM instance instance-00000023 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1032.783306] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb4805f-8c4d-47d1-9f87-338e863caf64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.808609] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] 9343301c-dfe9-41b0-b4a0-067af544d297/9343301c-dfe9-41b0-b4a0-067af544d297.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.809309] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6db8797-4a2f-494a-a8c3-463f876d6721 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.828926] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1032.828926] env[62814]: value = "task-4294258" [ 1032.828926] env[62814]: _type = "Task" [ 1032.828926] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.837615] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294258, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.947883] env[62814]: DEBUG oslo_vmware.api [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294254, 'name': ReconfigVM_Task, 'duration_secs': 1.277161} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.947883] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Reconfigured VM instance instance-00000035 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1032.952225] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f5ec808-8432-4adf-aa18-e54a3734e116 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.966822] env[62814]: DEBUG nova.scheduler.client.report [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1032.972192] env[62814]: DEBUG oslo_vmware.api [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1032.972192] env[62814]: value = "task-4294259" [ 1032.972192] env[62814]: _type = "Task" [ 1032.972192] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.984455] env[62814]: DEBUG oslo_vmware.api [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294259, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.071412] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4cd8da5f-9721-4be9-a78b-dfdd3cfa8647 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.268s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1033.234944] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294256, 'name': ReconfigVM_Task, 'duration_secs': 0.753842} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.235267] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 4ebc3280-64a7-4e9b-8385-2128357d3422/4ebc3280-64a7-4e9b-8385-2128357d3422.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.235913] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dbe60cdd-909e-41ec-9ec1-4ad1ac6ae0b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.242992] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1033.242992] env[62814]: value = "task-4294260" [ 1033.242992] env[62814]: _type = "Task" [ 1033.242992] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.251312] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294260, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.277114] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-a07b964f-fe88-4c88-b6f4-61ed0973716c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1033.277468] env[62814]: DEBUG nova.compute.manager [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Instance network_info: |[{"id": "08e04ba7-4665-4982-ac59-6412cc953d93", "address": "fa:16:3e:a3:d9:48", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08e04ba7-46", "ovs_interfaceid": "08e04ba7-4665-4982-ac59-6412cc953d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1033.277905] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:d9:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '605f83bd-808c-4b54-922e-54b14690987a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '08e04ba7-4665-4982-ac59-6412cc953d93', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1033.285617] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1033.285888] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1033.286137] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cec21ae1-4312-4074-bc46-2ca016a2ed7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.308176] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.308176] env[62814]: value = "task-4294261" [ 1033.308176] env[62814]: _type = "Task" [ 1033.308176] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.316140] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294261, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.339046] env[62814]: DEBUG oslo_vmware.api [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294258, 'name': ReconfigVM_Task, 'duration_secs': 0.288265} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.339188] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Reconfigured VM instance instance-00000023 to attach disk [datastore2] 9343301c-dfe9-41b0-b4a0-067af544d297/9343301c-dfe9-41b0-b4a0-067af544d297.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.339634] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance '9343301c-dfe9-41b0-b4a0-067af544d297' progress to 50 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1033.449970] env[62814]: DEBUG nova.compute.manager [req-7b549116-4e49-4a62-814c-bec7cdde709b req-d9855994-cd8f-47f7-ae3e-d27554fa53db service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Received event network-changed-08e04ba7-4665-4982-ac59-6412cc953d93 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1033.450185] env[62814]: DEBUG nova.compute.manager [req-7b549116-4e49-4a62-814c-bec7cdde709b req-d9855994-cd8f-47f7-ae3e-d27554fa53db service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Refreshing instance network info cache due to event network-changed-08e04ba7-4665-4982-ac59-6412cc953d93. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1033.450786] env[62814]: DEBUG oslo_concurrency.lockutils [req-7b549116-4e49-4a62-814c-bec7cdde709b req-d9855994-cd8f-47f7-ae3e-d27554fa53db service nova] Acquiring lock "refresh_cache-a07b964f-fe88-4c88-b6f4-61ed0973716c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.450942] env[62814]: DEBUG oslo_concurrency.lockutils [req-7b549116-4e49-4a62-814c-bec7cdde709b req-d9855994-cd8f-47f7-ae3e-d27554fa53db service nova] Acquired lock "refresh_cache-a07b964f-fe88-4c88-b6f4-61ed0973716c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1033.451121] env[62814]: DEBUG nova.network.neutron [req-7b549116-4e49-4a62-814c-bec7cdde709b req-d9855994-cd8f-47f7-ae3e-d27554fa53db service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Refreshing network info cache for port 08e04ba7-4665-4982-ac59-6412cc953d93 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1033.477878] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1033.478395] env[62814]: DEBUG nova.compute.manager [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1033.480940] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.559s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1033.482549] env[62814]: INFO nova.compute.claims [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.491477] env[62814]: DEBUG oslo_vmware.api [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294259, 'name': ReconfigVM_Task, 'duration_secs': 0.164414} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.491785] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845759', 'volume_id': 'f4d241a2-9ed8-46ae-af01-e7e67273236c', 'name': 'volume-f4d241a2-9ed8-46ae-af01-e7e67273236c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '9683894b-a300-4400-a1b9-db62478f42c5', 'attached_at': '', 'detached_at': '', 'volume_id': 'f4d241a2-9ed8-46ae-af01-e7e67273236c', 'serial': 'f4d241a2-9ed8-46ae-af01-e7e67273236c'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1033.575651] env[62814]: DEBUG nova.compute.manager [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1033.753328] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294260, 'name': Rename_Task, 'duration_secs': 0.142205} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.753623] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1033.753886] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8bd3159-4f5a-43fa-bc1c-166b9a43975e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.760893] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1033.760893] env[62814]: value = "task-4294262" [ 1033.760893] env[62814]: _type = "Task" [ 1033.760893] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.769198] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294262, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.818184] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294261, 'name': CreateVM_Task, 'duration_secs': 0.361939} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.818373] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1033.818953] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.819138] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1033.819525] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1033.819791] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fc105c0-5f7f-4e99-982a-9ce15cfaa9d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.825138] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1033.825138] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7bb8a-4fc1-1b34-5280-9826f28c4a57" [ 1033.825138] env[62814]: _type = "Task" [ 1033.825138] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.833711] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7bb8a-4fc1-1b34-5280-9826f28c4a57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.846222] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ad05bd-8aa6-4f2c-b9e8-db814031bde1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.866103] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd71304-c2f5-4434-b5f0-974fb2d9da69 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.886905] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance '9343301c-dfe9-41b0-b4a0-067af544d297' progress to 67 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1033.990540] env[62814]: DEBUG nova.compute.utils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1033.991707] env[62814]: DEBUG nova.compute.manager [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1033.991900] env[62814]: DEBUG nova.network.neutron [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1034.042102] env[62814]: DEBUG nova.objects.instance [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'flavor' on Instance uuid 9683894b-a300-4400-a1b9-db62478f42c5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.058786] env[62814]: DEBUG nova.policy [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b65daf2b92614e9eb97f247a7b07ef44', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'edccd256413a4a399377b6499c46e520', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1034.106161] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.265959] env[62814]: DEBUG nova.network.neutron [req-7b549116-4e49-4a62-814c-bec7cdde709b req-d9855994-cd8f-47f7-ae3e-d27554fa53db service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Updated VIF entry in instance network info cache for port 08e04ba7-4665-4982-ac59-6412cc953d93. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1034.266363] env[62814]: DEBUG nova.network.neutron [req-7b549116-4e49-4a62-814c-bec7cdde709b req-d9855994-cd8f-47f7-ae3e-d27554fa53db service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Updating instance_info_cache with network_info: [{"id": "08e04ba7-4665-4982-ac59-6412cc953d93", "address": "fa:16:3e:a3:d9:48", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08e04ba7-46", "ovs_interfaceid": "08e04ba7-4665-4982-ac59-6412cc953d93", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.283801] env[62814]: DEBUG oslo_vmware.api [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294262, 'name': PowerOnVM_Task, 'duration_secs': 0.500332} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.284081] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1034.284290] env[62814]: INFO nova.compute.manager [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Took 9.09 seconds to spawn the instance on the hypervisor. [ 1034.284469] env[62814]: DEBUG nova.compute.manager [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1034.285246] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a860df6d-b996-49a6-957b-288657221fdd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.330770] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "316931ae-7a62-4bac-81e4-1fee9a36164e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.330770] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1034.343646] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b7bb8a-4fc1-1b34-5280-9826f28c4a57, 'name': SearchDatastore_Task, 'duration_secs': 0.010584} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.344017] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1034.344376] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.344760] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.345089] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1034.345443] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.346113] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-854a848e-21fd-4acc-98c9-2bb78ec2430f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.357738] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.357952] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1034.359190] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c84348ed-dbf6-47ca-a351-80a215921424 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.365199] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1034.365199] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521f4935-1583-6775-a686-03367f25d84d" [ 1034.365199] env[62814]: _type = "Task" [ 1034.365199] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.374425] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521f4935-1583-6775-a686-03367f25d84d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.420919] env[62814]: DEBUG nova.network.neutron [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Successfully created port: 031fa2fc-70be-4d13-bb5d-8c516281f35e {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1034.429234] env[62814]: DEBUG nova.network.neutron [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Port 08bab1d2-6296-46f7-baf6-4344d1bbb0ef binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1034.495883] env[62814]: DEBUG nova.compute.manager [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1034.778109] env[62814]: DEBUG oslo_concurrency.lockutils [req-7b549116-4e49-4a62-814c-bec7cdde709b req-d9855994-cd8f-47f7-ae3e-d27554fa53db service nova] Releasing lock "refresh_cache-a07b964f-fe88-4c88-b6f4-61ed0973716c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1034.810515] env[62814]: INFO nova.compute.manager [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Took 33.92 seconds to build instance. [ 1034.833145] env[62814]: DEBUG nova.compute.utils [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1034.884024] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521f4935-1583-6775-a686-03367f25d84d, 'name': SearchDatastore_Task, 'duration_secs': 0.01271} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.884024] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e5db69f-b663-419e-813b-a74c72410245 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.887946] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1034.887946] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cd5d30-a837-af20-9fce-4e041f6dd797" [ 1034.887946] env[62814]: _type = "Task" [ 1034.887946] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.901429] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cd5d30-a837-af20-9fce-4e041f6dd797, 'name': SearchDatastore_Task, 'duration_secs': 0.010592} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.901429] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1034.901429] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a07b964f-fe88-4c88-b6f4-61ed0973716c/a07b964f-fe88-4c88-b6f4-61ed0973716c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1034.901783] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-78fe42f8-ef41-410e-935e-7ff65d4e4e20 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.911024] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1034.911024] env[62814]: value = "task-4294263" [ 1034.911024] env[62814]: _type = "Task" [ 1034.911024] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.919322] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294263, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.936696] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a556a76f-262b-4df7-b01e-310ac7d93a56 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.945016] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485895ed-5f9d-4c82-93d7-e80aae783192 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.976926] env[62814]: DEBUG oslo_concurrency.lockutils [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "9683894b-a300-4400-a1b9-db62478f42c5" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1034.977974] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb81b5c-1bb3-4aba-9552-c629e0012f57 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.986937] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80d57d0-5a57-4ce5-892f-548d00927aa9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.008232] env[62814]: DEBUG nova.compute.provider_tree [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.052600] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2deb561c-ca64-4845-a7c0-8025f3f96a91 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.325s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.053721] env[62814]: DEBUG oslo_concurrency.lockutils [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.077s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.053964] env[62814]: DEBUG nova.compute.manager [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1035.055009] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97fe4aca-a041-4bcf-864f-805679c118d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.063046] env[62814]: DEBUG nova.compute.manager [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1035.063275] env[62814]: DEBUG nova.objects.instance [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'flavor' on Instance uuid 9683894b-a300-4400-a1b9-db62478f42c5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1035.313213] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8090c5fd-4611-4de7-9127-6c475bd660da tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "4ebc3280-64a7-4e9b-8385-2128357d3422" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.441s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.339848] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.421178] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294263, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.454407] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "9343301c-dfe9-41b0-b4a0-067af544d297-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1035.454962] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1035.454962] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1035.510997] env[62814]: DEBUG nova.compute.manager [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1035.513722] env[62814]: DEBUG nova.scheduler.client.report [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1035.542169] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1035.542378] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1035.543053] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1035.543053] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1035.543053] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1035.543257] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1035.543333] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1035.543487] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1035.543672] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1035.543872] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1035.544098] env[62814]: DEBUG nova.virt.hardware [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1035.544976] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1046dbce-e63c-48a7-b321-cbd5bf9ac836 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.554387] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701ef2bd-68d7-43a6-bb80-e6ea50897df3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.816661] env[62814]: DEBUG nova.compute.manager [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1035.925388] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294263, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518532} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.925723] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a07b964f-fe88-4c88-b6f4-61ed0973716c/a07b964f-fe88-4c88-b6f4-61ed0973716c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1035.925984] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1035.926319] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-18c0e90b-37c9-4a5f-9b15-47bdff79942b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.935330] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1035.935330] env[62814]: value = "task-4294264" [ 1035.935330] env[62814]: _type = "Task" [ 1035.935330] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.946180] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294264, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.018930] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.019543] env[62814]: DEBUG nova.compute.manager [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1036.022752] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.497s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.024986] env[62814]: INFO nova.compute.claims [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.073610] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1036.073971] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00d9c51c-0178-458c-910b-a0ffc7b228eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.083397] env[62814]: DEBUG oslo_vmware.api [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1036.083397] env[62814]: value = "task-4294265" [ 1036.083397] env[62814]: _type = "Task" [ 1036.083397] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.095647] env[62814]: DEBUG oslo_vmware.api [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294265, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.175153] env[62814]: DEBUG nova.network.neutron [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Successfully updated port: 031fa2fc-70be-4d13-bb5d-8c516281f35e {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1036.233577] env[62814]: DEBUG nova.compute.manager [req-fc0c14dd-2ac9-4b0d-9c13-7d785d42207d req-d55d7b59-4f0f-4e56-bd31-7d10063b39b7 service nova] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Received event network-vif-plugged-031fa2fc-70be-4d13-bb5d-8c516281f35e {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1036.233797] env[62814]: DEBUG oslo_concurrency.lockutils [req-fc0c14dd-2ac9-4b0d-9c13-7d785d42207d req-d55d7b59-4f0f-4e56-bd31-7d10063b39b7 service nova] Acquiring lock "88ec5aba-f38d-4c75-af29-e3df3aa49640-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.233975] env[62814]: DEBUG oslo_concurrency.lockutils [req-fc0c14dd-2ac9-4b0d-9c13-7d785d42207d req-d55d7b59-4f0f-4e56-bd31-7d10063b39b7 service nova] Lock "88ec5aba-f38d-4c75-af29-e3df3aa49640-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.234191] env[62814]: DEBUG oslo_concurrency.lockutils [req-fc0c14dd-2ac9-4b0d-9c13-7d785d42207d req-d55d7b59-4f0f-4e56-bd31-7d10063b39b7 service nova] Lock "88ec5aba-f38d-4c75-af29-e3df3aa49640-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.234358] env[62814]: DEBUG nova.compute.manager [req-fc0c14dd-2ac9-4b0d-9c13-7d785d42207d req-d55d7b59-4f0f-4e56-bd31-7d10063b39b7 service nova] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] No waiting events found dispatching network-vif-plugged-031fa2fc-70be-4d13-bb5d-8c516281f35e {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1036.234525] env[62814]: WARNING nova.compute.manager [req-fc0c14dd-2ac9-4b0d-9c13-7d785d42207d req-d55d7b59-4f0f-4e56-bd31-7d10063b39b7 service nova] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Received unexpected event network-vif-plugged-031fa2fc-70be-4d13-bb5d-8c516281f35e for instance with vm_state building and task_state spawning. [ 1036.341401] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.419028] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "316931ae-7a62-4bac-81e4-1fee9a36164e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.419138] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.419331] env[62814]: INFO nova.compute.manager [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Attaching volume 4878e7fc-22eb-4976-bdda-753fce367d65 to /dev/sdb [ 1036.446072] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294264, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074736} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.446415] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1036.447472] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f21fd39d-8a41-4f9a-af7d-6a8582803b73 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.451359] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25f0a7d7-32ae-4fe2-a662-72b3eb8efb46 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.470365] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d9de47-08d3-4be0-80e8-191057ba556e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.481504] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] a07b964f-fe88-4c88-b6f4-61ed0973716c/a07b964f-fe88-4c88-b6f4-61ed0973716c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.482185] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6926e73a-6be8-4b13-a631-2211a8dcd6f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.503011] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1036.503011] env[62814]: value = "task-4294266" [ 1036.503011] env[62814]: _type = "Task" [ 1036.503011] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.514011] env[62814]: DEBUG nova.virt.block_device [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Updating existing volume attachment record: 0a71d74e-024c-4c56-a74c-15c606d76d98 {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1036.516476] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "4ebc3280-64a7-4e9b-8385-2128357d3422" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.516476] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "4ebc3280-64a7-4e9b-8385-2128357d3422" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.516476] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "4ebc3280-64a7-4e9b-8385-2128357d3422-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1036.516476] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "4ebc3280-64a7-4e9b-8385-2128357d3422-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1036.516476] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "4ebc3280-64a7-4e9b-8385-2128357d3422-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1036.519049] env[62814]: INFO nova.compute.manager [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Terminating instance [ 1036.524662] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294266, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.532308] env[62814]: DEBUG nova.compute.utils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1036.538352] env[62814]: DEBUG nova.compute.manager [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1036.538571] env[62814]: DEBUG nova.network.neutron [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1036.542919] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.543140] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1036.543361] env[62814]: DEBUG nova.network.neutron [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1036.597549] env[62814]: DEBUG oslo_vmware.api [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294265, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.610659] env[62814]: DEBUG nova.policy [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a07e756cf43641b3a56ba542a8b2f628', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e8761ca6d3444b2a94b18ef8628f4f4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1036.682794] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquiring lock "refresh_cache-88ec5aba-f38d-4c75-af29-e3df3aa49640" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.682794] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquired lock "refresh_cache-88ec5aba-f38d-4c75-af29-e3df3aa49640" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1036.682794] env[62814]: DEBUG nova.network.neutron [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1036.969430] env[62814]: DEBUG nova.network.neutron [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Successfully created port: 496dd017-79fb-4bcc-8fdb-32ab21f6b676 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.014508] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294266, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.027491] env[62814]: DEBUG nova.compute.manager [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1037.027857] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1037.028665] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed92fcb9-ef1e-4f2c-a06e-d8b2acaf73bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.040405] env[62814]: DEBUG nova.compute.manager [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1037.043130] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1037.043832] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ebddc0cd-8ec9-401c-ba65-bf5bd9436428 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.060689] env[62814]: DEBUG oslo_vmware.api [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1037.060689] env[62814]: value = "task-4294270" [ 1037.060689] env[62814]: _type = "Task" [ 1037.060689] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.073861] env[62814]: DEBUG oslo_vmware.api [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294270, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.106434] env[62814]: DEBUG oslo_vmware.api [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294265, 'name': PowerOffVM_Task, 'duration_secs': 0.594392} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.106434] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.106434] env[62814]: DEBUG nova.compute.manager [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1037.108783] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992c6d3e-d3e4-49bb-aeb9-4cbee7e2e9b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.267074] env[62814]: DEBUG nova.network.neutron [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1037.374918] env[62814]: DEBUG nova.network.neutron [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance_info_cache with network_info: [{"id": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "address": "fa:16:3e:db:ee:dc", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08bab1d2-62", "ovs_interfaceid": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.520202] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294266, 'name': ReconfigVM_Task, 'duration_secs': 0.577971} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.523636] env[62814]: DEBUG nova.network.neutron [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Updating instance_info_cache with network_info: [{"id": "031fa2fc-70be-4d13-bb5d-8c516281f35e", "address": "fa:16:3e:fd:1c:3f", "network": {"id": "9ad07879-150a-4b0f-8c7b-e1ae59f646b2", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-469903310-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "edccd256413a4a399377b6499c46e520", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap031fa2fc-70", "ovs_interfaceid": "031fa2fc-70be-4d13-bb5d-8c516281f35e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.524975] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Reconfigured VM instance instance-00000053 to attach disk [datastore2] a07b964f-fe88-4c88-b6f4-61ed0973716c/a07b964f-fe88-4c88-b6f4-61ed0973716c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.526303] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27a25ec9-55e9-4c2f-b6e8-16e1c806485e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.538635] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1037.538635] env[62814]: value = "task-4294271" [ 1037.538635] env[62814]: _type = "Task" [ 1037.538635] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.542045] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e76e34-4ef7-4e53-8c55-1daff9c352a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.556933] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294271, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.565286] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ca9eda-b755-4830-9902-1b0450adc483 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.576665] env[62814]: DEBUG oslo_vmware.api [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294270, 'name': PowerOffVM_Task, 'duration_secs': 0.230102} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.604504] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.604878] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1037.605931] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a03339b7-4a89-4cf3-91bb-0eaf9b4c19ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.608514] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf696f27-a925-4d8d-b0d8-7b0e818beb35 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.618851] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43a6cdd-55ff-4d75-ac79-862ed352687e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.639293] env[62814]: DEBUG nova.compute.provider_tree [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.641066] env[62814]: DEBUG oslo_concurrency.lockutils [None req-759d9f7d-3d82-4fd2-addc-9a2918adcc45 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.587s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1037.686782] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.687028] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.687216] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleting the datastore file [datastore2] 4ebc3280-64a7-4e9b-8385-2128357d3422 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.687944] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-77c7e2ae-4c21-45c2-93c3-d3ebaaa4a4a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.694951] env[62814]: DEBUG oslo_vmware.api [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1037.694951] env[62814]: value = "task-4294273" [ 1037.694951] env[62814]: _type = "Task" [ 1037.694951] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.703860] env[62814]: DEBUG oslo_vmware.api [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294273, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.880903] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.027508] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Releasing lock "refresh_cache-88ec5aba-f38d-4c75-af29-e3df3aa49640" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1038.027856] env[62814]: DEBUG nova.compute.manager [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Instance network_info: |[{"id": "031fa2fc-70be-4d13-bb5d-8c516281f35e", "address": "fa:16:3e:fd:1c:3f", "network": {"id": "9ad07879-150a-4b0f-8c7b-e1ae59f646b2", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-469903310-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "edccd256413a4a399377b6499c46e520", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap031fa2fc-70", "ovs_interfaceid": "031fa2fc-70be-4d13-bb5d-8c516281f35e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1038.028537] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fd:1c:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '140f4558-c11e-4af4-ab36-234e2d2f80a4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '031fa2fc-70be-4d13-bb5d-8c516281f35e', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1038.036245] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Creating folder: Project (edccd256413a4a399377b6499c46e520). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1038.036647] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-08572781-deea-485c-81d9-2cc6b76b24e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.047827] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294271, 'name': Rename_Task, 'duration_secs': 0.220593} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.048101] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1038.048350] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-def22ed2-32fd-47e5-ba35-545aed744760 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.054162] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Created folder: Project (edccd256413a4a399377b6499c46e520) in parent group-v845547. [ 1038.054431] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Creating folder: Instances. Parent ref: group-v845795. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1038.055662] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e3ce6724-2da6-470e-a196-8ba43fe13119 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.057260] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1038.057260] env[62814]: value = "task-4294275" [ 1038.057260] env[62814]: _type = "Task" [ 1038.057260] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.058380] env[62814]: DEBUG nova.compute.manager [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1038.069804] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294275, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.071225] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Created folder: Instances in parent group-v845795. [ 1038.071528] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1038.071727] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1038.071939] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62d3db29-d03a-42b9-b70b-d2582af668d8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.094330] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1038.094635] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.094757] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1038.094940] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.095120] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1038.095288] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1038.095486] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1038.095647] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1038.095813] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1038.095979] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1038.096163] env[62814]: DEBUG nova.virt.hardware [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1038.097024] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97d780b-4a2a-42d7-9318-d47e122c2e6f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.100486] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1038.100486] env[62814]: value = "task-4294277" [ 1038.100486] env[62814]: _type = "Task" [ 1038.100486] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.107693] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c79e633-3539-4fe0-903e-2c8e3d1348d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.114455] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294277, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.144142] env[62814]: DEBUG nova.scheduler.client.report [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1038.205842] env[62814]: DEBUG oslo_vmware.api [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294273, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.327322} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.206178] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.206368] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1038.206555] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.206740] env[62814]: INFO nova.compute.manager [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1038.206979] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1038.207186] env[62814]: DEBUG nova.compute.manager [-] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1038.207303] env[62814]: DEBUG nova.network.neutron [-] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1038.268924] env[62814]: DEBUG nova.compute.manager [req-beee6e93-e978-41e2-a802-de6be09eaf51 req-8b5733e8-b99c-4348-bd30-730f35e10847 service nova] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Received event network-changed-031fa2fc-70be-4d13-bb5d-8c516281f35e {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1038.269172] env[62814]: DEBUG nova.compute.manager [req-beee6e93-e978-41e2-a802-de6be09eaf51 req-8b5733e8-b99c-4348-bd30-730f35e10847 service nova] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Refreshing instance network info cache due to event network-changed-031fa2fc-70be-4d13-bb5d-8c516281f35e. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1038.269467] env[62814]: DEBUG oslo_concurrency.lockutils [req-beee6e93-e978-41e2-a802-de6be09eaf51 req-8b5733e8-b99c-4348-bd30-730f35e10847 service nova] Acquiring lock "refresh_cache-88ec5aba-f38d-4c75-af29-e3df3aa49640" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.269630] env[62814]: DEBUG oslo_concurrency.lockutils [req-beee6e93-e978-41e2-a802-de6be09eaf51 req-8b5733e8-b99c-4348-bd30-730f35e10847 service nova] Acquired lock "refresh_cache-88ec5aba-f38d-4c75-af29-e3df3aa49640" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1038.269848] env[62814]: DEBUG nova.network.neutron [req-beee6e93-e978-41e2-a802-de6be09eaf51 req-8b5733e8-b99c-4348-bd30-730f35e10847 service nova] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Refreshing network info cache for port 031fa2fc-70be-4d13-bb5d-8c516281f35e {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1038.411178] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e5ad8e-a6ff-4fb4-920d-585a2845d555 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.436360] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c0c633-9f43-486c-ba1b-e711a817f27b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.446253] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance '9343301c-dfe9-41b0-b4a0-067af544d297' progress to 83 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1038.570893] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294275, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.615472] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294277, 'name': CreateVM_Task, 'duration_secs': 0.505637} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.615472] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1038.615648] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.615817] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1038.616165] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1038.616429] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f68ee2ca-064b-49ec-9b9c-ec4708a60a2a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.627675] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for the task: (returnval){ [ 1038.627675] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520be3d7-1f5e-1fe3-e103-060114375930" [ 1038.627675] env[62814]: _type = "Task" [ 1038.627675] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.639737] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520be3d7-1f5e-1fe3-e103-060114375930, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.649610] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.627s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1038.650179] env[62814]: DEBUG nova.compute.manager [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1038.657017] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.627s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1038.657017] env[62814]: INFO nova.compute.claims [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1038.787991] env[62814]: DEBUG nova.compute.manager [req-8f2f9709-a81d-436d-98e6-551436828fa6 req-4d68546f-ddec-4232-a419-4f5916f64b14 service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Received event network-vif-deleted-af952a49-d433-42b8-8a05-fc3d6e780333 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1038.788212] env[62814]: INFO nova.compute.manager [req-8f2f9709-a81d-436d-98e6-551436828fa6 req-4d68546f-ddec-4232-a419-4f5916f64b14 service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Neutron deleted interface af952a49-d433-42b8-8a05-fc3d6e780333; detaching it from the instance and deleting it from the info cache [ 1038.788387] env[62814]: DEBUG nova.network.neutron [req-8f2f9709-a81d-436d-98e6-551436828fa6 req-4d68546f-ddec-4232-a419-4f5916f64b14 service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.955503] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a60d3a46-5770-4076-82e0-54e0a1c5299e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance '9343301c-dfe9-41b0-b4a0-067af544d297' progress to 100 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1039.015733] env[62814]: DEBUG nova.network.neutron [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Successfully updated port: 496dd017-79fb-4bcc-8fdb-32ab21f6b676 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1039.071723] env[62814]: DEBUG oslo_vmware.api [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294275, 'name': PowerOnVM_Task, 'duration_secs': 0.696284} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.072124] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1039.072335] env[62814]: INFO nova.compute.manager [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Took 8.17 seconds to spawn the instance on the hypervisor. [ 1039.072638] env[62814]: DEBUG nova.compute.manager [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1039.073522] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa116965-5c4a-4771-b4be-9724ef30803c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.132707] env[62814]: DEBUG nova.network.neutron [req-beee6e93-e978-41e2-a802-de6be09eaf51 req-8b5733e8-b99c-4348-bd30-730f35e10847 service nova] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Updated VIF entry in instance network info cache for port 031fa2fc-70be-4d13-bb5d-8c516281f35e. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1039.133105] env[62814]: DEBUG nova.network.neutron [req-beee6e93-e978-41e2-a802-de6be09eaf51 req-8b5733e8-b99c-4348-bd30-730f35e10847 service nova] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Updating instance_info_cache with network_info: [{"id": "031fa2fc-70be-4d13-bb5d-8c516281f35e", "address": "fa:16:3e:fd:1c:3f", "network": {"id": "9ad07879-150a-4b0f-8c7b-e1ae59f646b2", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-469903310-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "edccd256413a4a399377b6499c46e520", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "140f4558-c11e-4af4-ab36-234e2d2f80a4", "external-id": "nsx-vlan-transportzone-638", "segmentation_id": 638, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap031fa2fc-70", "ovs_interfaceid": "031fa2fc-70be-4d13-bb5d-8c516281f35e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.135110] env[62814]: DEBUG nova.network.neutron [-] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.143863] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520be3d7-1f5e-1fe3-e103-060114375930, 'name': SearchDatastore_Task, 'duration_secs': 0.030307} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.144477] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1039.144718] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1039.144968] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.145135] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1039.145311] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1039.146406] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-71d05171-ea92-4a16-8184-2d3f32e4550a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.156838] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1039.156978] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1039.157934] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b4336da-e752-4b5f-95e6-94e85d84f34f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.162838] env[62814]: DEBUG nova.compute.utils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1039.169023] env[62814]: DEBUG nova.compute.manager [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1039.169023] env[62814]: DEBUG nova.network.neutron [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1039.173060] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for the task: (returnval){ [ 1039.173060] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527099fa-22ac-1000-6f28-d485244e8bc3" [ 1039.173060] env[62814]: _type = "Task" [ 1039.173060] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.184527] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527099fa-22ac-1000-6f28-d485244e8bc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.206651] env[62814]: DEBUG nova.policy [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a07e756cf43641b3a56ba542a8b2f628', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e8761ca6d3444b2a94b18ef8628f4f4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1039.224618] env[62814]: DEBUG nova.objects.instance [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'flavor' on Instance uuid 9683894b-a300-4400-a1b9-db62478f42c5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1039.290655] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6604fe6d-d15c-44b4-ba00-76c950abaa0d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.301164] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73de026-b2b3-41e9-9557-5306829dfc93 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.336713] env[62814]: DEBUG nova.compute.manager [req-8f2f9709-a81d-436d-98e6-551436828fa6 req-4d68546f-ddec-4232-a419-4f5916f64b14 service nova] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Detach interface failed, port_id=af952a49-d433-42b8-8a05-fc3d6e780333, reason: Instance 4ebc3280-64a7-4e9b-8385-2128357d3422 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1039.439071] env[62814]: DEBUG nova.network.neutron [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Successfully created port: 76bd40fc-ecff-43a9-a0e9-f58b9500f9c5 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.518563] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "refresh_cache-cfab341e-57a1-48b4-9b6b-ceecf28c223b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.518784] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired lock "refresh_cache-cfab341e-57a1-48b4-9b6b-ceecf28c223b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1039.518966] env[62814]: DEBUG nova.network.neutron [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1039.594512] env[62814]: INFO nova.compute.manager [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Took 33.95 seconds to build instance. [ 1039.636736] env[62814]: DEBUG oslo_concurrency.lockutils [req-beee6e93-e978-41e2-a802-de6be09eaf51 req-8b5733e8-b99c-4348-bd30-730f35e10847 service nova] Releasing lock "refresh_cache-88ec5aba-f38d-4c75-af29-e3df3aa49640" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1039.644440] env[62814]: INFO nova.compute.manager [-] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Took 1.44 seconds to deallocate network for instance. [ 1039.668379] env[62814]: DEBUG nova.compute.manager [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1039.686061] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527099fa-22ac-1000-6f28-d485244e8bc3, 'name': SearchDatastore_Task, 'duration_secs': 0.020653} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.686860] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4d5781d-67f9-417c-82d3-dcdbb597ae27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.695261] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for the task: (returnval){ [ 1039.695261] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5220355c-ff12-d738-8454-bf158f096f44" [ 1039.695261] env[62814]: _type = "Task" [ 1039.695261] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.710049] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5220355c-ff12-d738-8454-bf158f096f44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.735023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.735023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquired lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1039.735023] env[62814]: DEBUG nova.network.neutron [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1039.735023] env[62814]: DEBUG nova.objects.instance [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'info_cache' on Instance uuid 9683894b-a300-4400-a1b9-db62478f42c5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1040.059279] env[62814]: DEBUG nova.network.neutron [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1040.096939] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df4b421-d67d-4e68-bdbb-de8d788b7480 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "a07b964f-fe88-4c88-b6f4-61ed0973716c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.467s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.103055] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e995dc-81df-4d74-ba7a-1ae0780b1e84 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.114237] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a1e46f-1cef-4e98-a0c3-aa4a5329a560 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.146186] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52c9bf2-54b5-433f-b509-60e93bf6523a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.152529] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.157230] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3626dcdd-d59f-4a09-8805-e3118c9cd9f7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.170572] env[62814]: DEBUG nova.compute.provider_tree [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.211977] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5220355c-ff12-d738-8454-bf158f096f44, 'name': SearchDatastore_Task, 'duration_secs': 0.027513} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.212503] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1040.212760] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 88ec5aba-f38d-4c75-af29-e3df3aa49640/88ec5aba-f38d-4c75-af29-e3df3aa49640.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1040.213034] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6d57d891-7c08-4dce-a55c-b3357ac12083 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.222232] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for the task: (returnval){ [ 1040.222232] env[62814]: value = "task-4294279" [ 1040.222232] env[62814]: _type = "Task" [ 1040.222232] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.231896] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294279, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.236215] env[62814]: DEBUG nova.objects.base [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Object Instance<9683894b-a300-4400-a1b9-db62478f42c5> lazy-loaded attributes: flavor,info_cache {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1040.239234] env[62814]: DEBUG nova.network.neutron [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Updating instance_info_cache with network_info: [{"id": "496dd017-79fb-4bcc-8fdb-32ab21f6b676", "address": "fa:16:3e:22:4b:29", "network": {"id": "c3a9b4ad-b825-451e-ad5a-4e0331bb2d16", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1261267397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e8761ca6d3444b2a94b18ef8628f4f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap496dd017-79", "ovs_interfaceid": "496dd017-79fb-4bcc-8fdb-32ab21f6b676", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.673789] env[62814]: DEBUG nova.scheduler.client.report [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1040.680764] env[62814]: DEBUG nova.compute.manager [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1040.708147] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1040.708405] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.708566] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1040.708800] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.708984] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1040.709155] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1040.709371] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1040.709531] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1040.709700] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1040.709862] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1040.710043] env[62814]: DEBUG nova.virt.hardware [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1040.710988] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cfb8c8-009f-491d-b489-b160420da23a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.713929] env[62814]: INFO nova.compute.manager [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Rebuilding instance [ 1040.722246] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e721450a-13db-43a4-ba34-9dd9e0095d41 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.748205] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Releasing lock "refresh_cache-cfab341e-57a1-48b4-9b6b-ceecf28c223b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1040.748560] env[62814]: DEBUG nova.compute.manager [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Instance network_info: |[{"id": "496dd017-79fb-4bcc-8fdb-32ab21f6b676", "address": "fa:16:3e:22:4b:29", "network": {"id": "c3a9b4ad-b825-451e-ad5a-4e0331bb2d16", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1261267397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e8761ca6d3444b2a94b18ef8628f4f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap496dd017-79", "ovs_interfaceid": "496dd017-79fb-4bcc-8fdb-32ab21f6b676", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1040.748815] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294279, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490649} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.749368] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:22:4b:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '496dd017-79fb-4bcc-8fdb-32ab21f6b676', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1040.756837] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Creating folder: Project (2e8761ca6d3444b2a94b18ef8628f4f4). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1040.757957] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 88ec5aba-f38d-4c75-af29-e3df3aa49640/88ec5aba-f38d-4c75-af29-e3df3aa49640.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1040.757957] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1040.757957] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01a585f5-dc89-4e4a-ad24-16d2a3a66a17 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.759468] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1cfe19fa-838a-4cc9-baab-d3796cffe8c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.767759] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for the task: (returnval){ [ 1040.767759] env[62814]: value = "task-4294281" [ 1040.767759] env[62814]: _type = "Task" [ 1040.767759] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.768152] env[62814]: DEBUG nova.compute.manager [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1040.768947] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4dfda38-63ac-4fc9-898c-6a1b494f8897 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.776319] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Created folder: Project (2e8761ca6d3444b2a94b18ef8628f4f4) in parent group-v845547. [ 1040.776466] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Creating folder: Instances. Parent ref: group-v845798. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1040.777066] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a41f09b-1fd8-4f48-a1cb-2505fb2052e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.786825] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294281, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.796661] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Created folder: Instances in parent group-v845798. [ 1040.796984] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1040.797356] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1040.797541] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04280a6e-3cf9-4d96-b45e-f50f7c1ab612 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.820532] env[62814]: DEBUG nova.compute.manager [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Received event network-vif-plugged-496dd017-79fb-4bcc-8fdb-32ab21f6b676 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1040.820834] env[62814]: DEBUG oslo_concurrency.lockutils [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] Acquiring lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1040.821080] env[62814]: DEBUG oslo_concurrency.lockutils [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] Lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1040.821284] env[62814]: DEBUG oslo_concurrency.lockutils [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] Lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1040.821569] env[62814]: DEBUG nova.compute.manager [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] No waiting events found dispatching network-vif-plugged-496dd017-79fb-4bcc-8fdb-32ab21f6b676 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1040.821786] env[62814]: WARNING nova.compute.manager [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Received unexpected event network-vif-plugged-496dd017-79fb-4bcc-8fdb-32ab21f6b676 for instance with vm_state building and task_state spawning. [ 1040.822126] env[62814]: DEBUG nova.compute.manager [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Received event network-changed-496dd017-79fb-4bcc-8fdb-32ab21f6b676 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1040.822399] env[62814]: DEBUG nova.compute.manager [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Refreshing instance network info cache due to event network-changed-496dd017-79fb-4bcc-8fdb-32ab21f6b676. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1040.822594] env[62814]: DEBUG oslo_concurrency.lockutils [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] Acquiring lock "refresh_cache-cfab341e-57a1-48b4-9b6b-ceecf28c223b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.822739] env[62814]: DEBUG oslo_concurrency.lockutils [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] Acquired lock "refresh_cache-cfab341e-57a1-48b4-9b6b-ceecf28c223b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1040.822894] env[62814]: DEBUG nova.network.neutron [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Refreshing network info cache for port 496dd017-79fb-4bcc-8fdb-32ab21f6b676 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1040.826620] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1040.826620] env[62814]: value = "task-4294283" [ 1040.826620] env[62814]: _type = "Task" [ 1040.826620] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.838842] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294283, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.065202] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1041.065465] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845794', 'volume_id': '4878e7fc-22eb-4976-bdda-753fce367d65', 'name': 'volume-4878e7fc-22eb-4976-bdda-753fce367d65', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '316931ae-7a62-4bac-81e4-1fee9a36164e', 'attached_at': '', 'detached_at': '', 'volume_id': '4878e7fc-22eb-4976-bdda-753fce367d65', 'serial': '4878e7fc-22eb-4976-bdda-753fce367d65'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1041.066377] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7563dee5-7cbc-45c7-bd76-9da1144bb0e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.087318] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3f4e59-d31e-4f5e-b5a4-165d7041205e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.122992] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] volume-4878e7fc-22eb-4976-bdda-753fce367d65/volume-4878e7fc-22eb-4976-bdda-753fce367d65.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.124033] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12a715c7-2eb2-4ec3-8e8a-409fac511c1c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.148605] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1041.148605] env[62814]: value = "task-4294284" [ 1041.148605] env[62814]: _type = "Task" [ 1041.148605] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.155241] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294284, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.179788] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1041.180345] env[62814]: DEBUG nova.compute.manager [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1041.185865] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 27.660s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1041.206895] env[62814]: DEBUG nova.network.neutron [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Updating instance_info_cache with network_info: [{"id": "9fed46ec-04e6-4482-a460-370e4ff04a32", "address": "fa:16:3e:19:a2:ac", "network": {"id": "5363651c-6ade-4a4c-b762-9fc7cbbcf5ea", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1833251571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b3ba1c9c7624abfb3e0bd880dc40166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fed46ec-04", "ovs_interfaceid": "9fed46ec-04e6-4482-a460-370e4ff04a32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.286187] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294281, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.126076} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.286532] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1041.287423] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae9b0ca-23ab-4ccf-aa79-d100c40f4250 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.315071] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 88ec5aba-f38d-4c75-af29-e3df3aa49640/88ec5aba-f38d-4c75-af29-e3df3aa49640.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1041.315869] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99f4f3e7-578f-4def-b05b-d24ecda340bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.343272] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294283, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.344991] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for the task: (returnval){ [ 1041.344991] env[62814]: value = "task-4294285" [ 1041.344991] env[62814]: _type = "Task" [ 1041.344991] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.353854] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294285, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.453287] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "9343301c-dfe9-41b0-b4a0-067af544d297" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1041.453894] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1041.454114] env[62814]: DEBUG nova.compute.manager [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Going to confirm migration 3 {{(pid=62814) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 1041.582301] env[62814]: DEBUG nova.network.neutron [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Updated VIF entry in instance network info cache for port 496dd017-79fb-4bcc-8fdb-32ab21f6b676. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1041.582685] env[62814]: DEBUG nova.network.neutron [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Updating instance_info_cache with network_info: [{"id": "496dd017-79fb-4bcc-8fdb-32ab21f6b676", "address": "fa:16:3e:22:4b:29", "network": {"id": "c3a9b4ad-b825-451e-ad5a-4e0331bb2d16", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1261267397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e8761ca6d3444b2a94b18ef8628f4f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap496dd017-79", "ovs_interfaceid": "496dd017-79fb-4bcc-8fdb-32ab21f6b676", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.656964] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.709505] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Releasing lock "refresh_cache-9683894b-a300-4400-a1b9-db62478f42c5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1041.794674] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1041.794939] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e54c85e6-22ad-4ade-af69-13c7a05316d8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.804035] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1041.804035] env[62814]: value = "task-4294286" [ 1041.804035] env[62814]: _type = "Task" [ 1041.804035] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.814419] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.841941] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294283, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.855551] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294285, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.018179] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.018374] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1042.018551] env[62814]: DEBUG nova.network.neutron [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1042.018736] env[62814]: DEBUG nova.objects.instance [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'info_cache' on Instance uuid 9343301c-dfe9-41b0-b4a0-067af544d297 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.085686] env[62814]: DEBUG oslo_concurrency.lockutils [req-fa5c3022-9776-4abe-bac0-9b2f25367ac0 req-2f10d577-44b1-4e47-a8f7-3fd5fafc367a service nova] Releasing lock "refresh_cache-cfab341e-57a1-48b4-9b6b-ceecf28c223b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1042.157987] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.201018] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Applying migration context for instance 9343301c-dfe9-41b0-b4a0-067af544d297 as it has an incoming, in-progress migration 06826230-b161-48d6-bc89-eabd48b17ffe. Migration status is confirming {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1042.201018] env[62814]: INFO nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating resource usage from migration 953ceb03-e52d-4c0a-9b6d-70bce2963e8d [ 1042.201018] env[62814]: INFO nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating resource usage from migration 06826230-b161-48d6-bc89-eabd48b17ffe [ 1042.225170] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance c7cef7f3-11db-44e1-a454-98830b465b52 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.225349] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d9242042-6209-4b04-bf00-00dd04d9d6a0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.225476] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 9683894b-a300-4400-a1b9-db62478f42c5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.225594] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 1e259ec6-d31a-453b-87e0-baa446665d56 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.225716] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 7350d352-9336-40b8-81a6-0a4795d9f8dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.225824] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance f54b8bb9-69b7-4bb4-a82c-9f796050e719 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.225951] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 508321ab-ce10-4953-a9e3-193b9975bec7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.226091] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 40e2d845-0211-4c84-aef7-94014f999e1d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.226218] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance fd1867e3-2523-4969-a157-b14c650f3779 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.226334] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 885e0798-2fa6-4f6b-82be-517b6d1168d2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.226456] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance a68901a9-9a9b-4127-bca9-64f98dfb151f is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.226570] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance e727eeb1-c5d0-4591-80bb-31746bf976a2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.226682] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance edcdb1e5-ed9d-49e6-97e6-ea7629682547 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.226791] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 316931ae-7a62-4bac-81e4-1fee9a36164e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.226910] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 4ebc3280-64a7-4e9b-8385-2128357d3422 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1042.227031] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Migration 06826230-b161-48d6-bc89-eabd48b17ffe is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1042.227148] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 9343301c-dfe9-41b0-b4a0-067af544d297 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.227257] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance a07b964f-fe88-4c88-b6f4-61ed0973716c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.227365] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 88ec5aba-f38d-4c75-af29-e3df3aa49640 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.227473] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance cfab341e-57a1-48b4-9b6b-ceecf28c223b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.227608] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 1c86646f-6f38-4f8d-bea0-8a6b93680aba actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.227748] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance b4ee4cee-3298-4955-8375-8ca8c04b2f9f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.227862] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Migration 953ceb03-e52d-4c0a-9b6d-70bce2963e8d is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1042.227971] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 59d21ef7-df97-49ac-9329-4c18df6dd087 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1042.315481] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.342713] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294283, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.354041] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294285, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.659415] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.715249] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1042.715697] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f94d35c3-ea0d-433b-b4b0-ecc6b6b48269 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.724814] env[62814]: DEBUG oslo_vmware.api [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1042.724814] env[62814]: value = "task-4294287" [ 1042.724814] env[62814]: _type = "Task" [ 1042.724814] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.734132] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 0c6c598f-ef5a-4e91-b811-cd3d8f072647 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1042.735440] env[62814]: DEBUG oslo_vmware.api [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294287, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.816062] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.843396] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294283, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.855693] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294285, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.881827] env[62814]: DEBUG nova.compute.manager [req-856355e2-e797-42ee-a38b-278cc788fe76 req-4d8dce39-e355-4cf2-96d5-7c3d403bea8d service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Received event network-vif-plugged-76bd40fc-ecff-43a9-a0e9-f58b9500f9c5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1042.881827] env[62814]: DEBUG oslo_concurrency.lockutils [req-856355e2-e797-42ee-a38b-278cc788fe76 req-4d8dce39-e355-4cf2-96d5-7c3d403bea8d service nova] Acquiring lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1042.881827] env[62814]: DEBUG oslo_concurrency.lockutils [req-856355e2-e797-42ee-a38b-278cc788fe76 req-4d8dce39-e355-4cf2-96d5-7c3d403bea8d service nova] Lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1042.881827] env[62814]: DEBUG oslo_concurrency.lockutils [req-856355e2-e797-42ee-a38b-278cc788fe76 req-4d8dce39-e355-4cf2-96d5-7c3d403bea8d service nova] Lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1042.882511] env[62814]: DEBUG nova.compute.manager [req-856355e2-e797-42ee-a38b-278cc788fe76 req-4d8dce39-e355-4cf2-96d5-7c3d403bea8d service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] No waiting events found dispatching network-vif-plugged-76bd40fc-ecff-43a9-a0e9-f58b9500f9c5 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1042.882511] env[62814]: WARNING nova.compute.manager [req-856355e2-e797-42ee-a38b-278cc788fe76 req-4d8dce39-e355-4cf2-96d5-7c3d403bea8d service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Received unexpected event network-vif-plugged-76bd40fc-ecff-43a9-a0e9-f58b9500f9c5 for instance with vm_state building and task_state spawning. [ 1043.158230] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.219519] env[62814]: DEBUG nova.network.neutron [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance_info_cache with network_info: [{"id": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "address": "fa:16:3e:db:ee:dc", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08bab1d2-62", "ovs_interfaceid": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.236013] env[62814]: DEBUG oslo_vmware.api [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294287, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.236689] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance cd037f6e-fddd-4389-b6d0-144b798537bf has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1043.316635] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.344090] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294283, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.355672] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294285, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.559518] env[62814]: DEBUG nova.network.neutron [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Successfully updated port: 76bd40fc-ecff-43a9-a0e9-f58b9500f9c5 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1043.659906] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294284, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.690778] env[62814]: DEBUG nova.compute.utils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1043.693263] env[62814]: DEBUG nova.compute.manager [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1043.693490] env[62814]: DEBUG nova.network.neutron [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1043.723217] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1043.723425] env[62814]: DEBUG nova.objects.instance [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'migration_context' on Instance uuid 9343301c-dfe9-41b0-b4a0-067af544d297 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.736902] env[62814]: DEBUG oslo_vmware.api [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294287, 'name': PowerOnVM_Task, 'duration_secs': 0.896547} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.737182] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1043.737375] env[62814]: DEBUG nova.compute.manager [None req-8f4976ab-dcc7-4d6c-8b66-f98ef42e014c tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1043.738223] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e7a44a-88ea-4e7b-9f16-55f64490781b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.741873] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1043.760889] env[62814]: DEBUG nova.policy [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a07e756cf43641b3a56ba542a8b2f628', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e8761ca6d3444b2a94b18ef8628f4f4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1043.817437] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294286, 'name': PowerOffVM_Task, 'duration_secs': 1.8673} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.817726] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1043.817989] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1043.819051] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02450506-ce99-4a31-93da-c5cd6d2b9bff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.826498] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1043.826729] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0563edd3-c257-4ffe-a101-2a54b4622da3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.843676] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294283, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.856290] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294285, 'name': ReconfigVM_Task, 'duration_secs': 2.359508} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.856584] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 88ec5aba-f38d-4c75-af29-e3df3aa49640/88ec5aba-f38d-4c75-af29-e3df3aa49640.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.857239] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81e46fb8-395c-44b2-9fbb-4ac10647fec3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.864179] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for the task: (returnval){ [ 1043.864179] env[62814]: value = "task-4294289" [ 1043.864179] env[62814]: _type = "Task" [ 1043.864179] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.872629] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294289, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.985053] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "interface-885e0798-2fa6-4f6b-82be-517b6d1168d2-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1043.986897] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-885e0798-2fa6-4f6b-82be-517b6d1168d2-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1043.986897] env[62814]: DEBUG nova.objects.instance [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'flavor' on Instance uuid 885e0798-2fa6-4f6b-82be-517b6d1168d2 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.055362] env[62814]: DEBUG nova.network.neutron [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Successfully created port: 89bae162-2360-4722-8b1c-ade210004fcc {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1044.062142] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "refresh_cache-1c86646f-6f38-4f8d-bea0-8a6b93680aba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.062259] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired lock "refresh_cache-1c86646f-6f38-4f8d-bea0-8a6b93680aba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1044.062480] env[62814]: DEBUG nova.network.neutron [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1044.159843] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294284, 'name': ReconfigVM_Task, 'duration_secs': 2.608031} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.160236] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Reconfigured VM instance instance-00000052 to attach disk [datastore1] volume-4878e7fc-22eb-4976-bdda-753fce367d65/volume-4878e7fc-22eb-4976-bdda-753fce367d65.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1044.165199] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2e2fdd57-dfff-4e23-aac7-6bd4005d4246 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.182475] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1044.182475] env[62814]: value = "task-4294290" [ 1044.182475] env[62814]: _type = "Task" [ 1044.182475] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.195171] env[62814]: DEBUG nova.compute.manager [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1044.199031] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294290, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.227362] env[62814]: DEBUG nova.objects.base [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Object Instance<9343301c-dfe9-41b0-b4a0-067af544d297> lazy-loaded attributes: info_cache,migration_context {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1044.228352] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50b41d6-73b1-4499-b91e-aab9bc7413f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.248691] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d5df1d0e-55e6-452e-882a-053a83250b39 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1044.253234] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00e014bd-55fb-433c-a21a-dcd8456f5620 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.260248] env[62814]: DEBUG oslo_vmware.api [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1044.260248] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528af608-e125-fe75-87d6-9118fbce3181" [ 1044.260248] env[62814]: _type = "Task" [ 1044.260248] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.274893] env[62814]: DEBUG oslo_vmware.api [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528af608-e125-fe75-87d6-9118fbce3181, 'name': SearchDatastore_Task, 'duration_secs': 0.010465} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.275179] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1044.346118] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294283, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.374877] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294289, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.483652] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1044.484531] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1044.484531] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleting the datastore file [datastore2] a07b964f-fe88-4c88-b6f4-61ed0973716c {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1044.485277] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a73c30a6-20e6-4823-99c2-0474ff3b1d9c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.495953] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1044.495953] env[62814]: value = "task-4294291" [ 1044.495953] env[62814]: _type = "Task" [ 1044.495953] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.509406] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.583501] env[62814]: DEBUG nova.objects.instance [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'pci_requests' on Instance uuid 885e0798-2fa6-4f6b-82be-517b6d1168d2 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.616160] env[62814]: DEBUG nova.network.neutron [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1044.691836] env[62814]: DEBUG oslo_vmware.api [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294290, 'name': ReconfigVM_Task, 'duration_secs': 0.40054} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.692169] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845794', 'volume_id': '4878e7fc-22eb-4976-bdda-753fce367d65', 'name': 'volume-4878e7fc-22eb-4976-bdda-753fce367d65', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '316931ae-7a62-4bac-81e4-1fee9a36164e', 'attached_at': '', 'detached_at': '', 'volume_id': '4878e7fc-22eb-4976-bdda-753fce367d65', 'serial': '4878e7fc-22eb-4976-bdda-753fce367d65'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1044.755821] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 44ea319c-6ea0-456a-bee6-42133a25d8c1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1044.756133] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 17 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1044.756285] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3904MB phys_disk=149GB used_disk=17GB total_vcpus=48 used_vcpus=17 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1044.800504] env[62814]: DEBUG nova.network.neutron [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Updating instance_info_cache with network_info: [{"id": "76bd40fc-ecff-43a9-a0e9-f58b9500f9c5", "address": "fa:16:3e:f8:d9:df", "network": {"id": "c3a9b4ad-b825-451e-ad5a-4e0331bb2d16", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1261267397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e8761ca6d3444b2a94b18ef8628f4f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76bd40fc-ec", "ovs_interfaceid": "76bd40fc-ecff-43a9-a0e9-f58b9500f9c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.850501] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294283, 'name': CreateVM_Task, 'duration_secs': 3.656151} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.850658] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1044.851362] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1044.851587] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1044.851925] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1044.854411] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b99ebe02-a243-4713-9357-8abf47d120fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.861852] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1044.861852] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521b4ed0-7aeb-fa48-1334-cfd8cccb6c4c" [ 1044.861852] env[62814]: _type = "Task" [ 1044.861852] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.868290] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521b4ed0-7aeb-fa48-1334-cfd8cccb6c4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.878460] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294289, 'name': Rename_Task, 'duration_secs': 0.738459} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.878953] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1044.878953] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ed64a9b-23b3-4f87-8cbe-bfb99f0c3216 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.887219] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for the task: (returnval){ [ 1044.887219] env[62814]: value = "task-4294292" [ 1044.887219] env[62814]: _type = "Task" [ 1044.887219] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.895413] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294292, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.007311] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184245} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.007926] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1045.007926] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1045.007926] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1045.086127] env[62814]: DEBUG nova.objects.base [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Object Instance<885e0798-2fa6-4f6b-82be-517b6d1168d2> lazy-loaded attributes: flavor,pci_requests {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1045.086375] env[62814]: DEBUG nova.network.neutron [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1045.120353] env[62814]: DEBUG nova.policy [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5779eaa60ec44b2b80660e521b598eb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10018060297c409d82935ad626cc54c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1045.133209] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0cc1198-4024-4b7e-9608-0474bcf904ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.141776] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a91c3c2-241a-421c-bb49-cd41bfc7f2ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.174677] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-637ba43f-6ae3-40a6-8acd-320057b7d0fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.183376] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd5d0db-6cee-4949-89c7-966bfd4d020f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.199998] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.207884] env[62814]: DEBUG nova.compute.manager [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1045.224072] env[62814]: DEBUG nova.compute.manager [req-c43ce6f7-dd82-49d8-b528-1c4db6fc7c1c req-6c24ec83-8906-4196-a005-eecbf52ce0df service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Received event network-changed-76bd40fc-ecff-43a9-a0e9-f58b9500f9c5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1045.224307] env[62814]: DEBUG nova.compute.manager [req-c43ce6f7-dd82-49d8-b528-1c4db6fc7c1c req-6c24ec83-8906-4196-a005-eecbf52ce0df service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Refreshing instance network info cache due to event network-changed-76bd40fc-ecff-43a9-a0e9-f58b9500f9c5. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1045.224511] env[62814]: DEBUG oslo_concurrency.lockutils [req-c43ce6f7-dd82-49d8-b528-1c4db6fc7c1c req-6c24ec83-8906-4196-a005-eecbf52ce0df service nova] Acquiring lock "refresh_cache-1c86646f-6f38-4f8d-bea0-8a6b93680aba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.236914] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1045.237157] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.237339] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1045.237547] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.237710] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1045.237854] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1045.238103] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1045.238281] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1045.239641] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1045.239641] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1045.239641] env[62814]: DEBUG nova.virt.hardware [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1045.239641] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe47140-4d20-4474-8e5a-bd9fdef294b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.250404] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76bb4a9-e846-4d37-9cca-6e04b5d6a9d6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.303548] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Releasing lock "refresh_cache-1c86646f-6f38-4f8d-bea0-8a6b93680aba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1045.303904] env[62814]: DEBUG nova.compute.manager [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Instance network_info: |[{"id": "76bd40fc-ecff-43a9-a0e9-f58b9500f9c5", "address": "fa:16:3e:f8:d9:df", "network": {"id": "c3a9b4ad-b825-451e-ad5a-4e0331bb2d16", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1261267397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e8761ca6d3444b2a94b18ef8628f4f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76bd40fc-ec", "ovs_interfaceid": "76bd40fc-ecff-43a9-a0e9-f58b9500f9c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1045.304260] env[62814]: DEBUG oslo_concurrency.lockutils [req-c43ce6f7-dd82-49d8-b528-1c4db6fc7c1c req-6c24ec83-8906-4196-a005-eecbf52ce0df service nova] Acquired lock "refresh_cache-1c86646f-6f38-4f8d-bea0-8a6b93680aba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.304445] env[62814]: DEBUG nova.network.neutron [req-c43ce6f7-dd82-49d8-b528-1c4db6fc7c1c req-6c24ec83-8906-4196-a005-eecbf52ce0df service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Refreshing network info cache for port 76bd40fc-ecff-43a9-a0e9-f58b9500f9c5 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1045.305672] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:d9:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '76bd40fc-ecff-43a9-a0e9-f58b9500f9c5', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1045.314213] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1045.317730] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1045.318249] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b80e8e4-3242-4e0b-85f2-da112f0186a2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.348421] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1045.348421] env[62814]: value = "task-4294293" [ 1045.348421] env[62814]: _type = "Task" [ 1045.348421] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.358776] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294293, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.373968] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521b4ed0-7aeb-fa48-1334-cfd8cccb6c4c, 'name': SearchDatastore_Task, 'duration_secs': 0.011864} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.374340] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1045.374614] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1045.374846] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.375018] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.375257] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1045.375492] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24811286-48fd-4952-86d9-40f355928558 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.395534] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1045.395741] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1045.396918] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4fe44204-64dc-4033-8c92-724545006b51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.403349] env[62814]: DEBUG oslo_vmware.api [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294292, 'name': PowerOnVM_Task, 'duration_secs': 0.502975} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.404013] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1045.404254] env[62814]: INFO nova.compute.manager [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Took 9.89 seconds to spawn the instance on the hypervisor. [ 1045.404440] env[62814]: DEBUG nova.compute.manager [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1045.405212] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c51812-4a1c-497d-913d-2f61d4683613 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.409838] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1045.409838] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c185bd-8a93-bbe3-4f49-6ee770a89337" [ 1045.409838] env[62814]: _type = "Task" [ 1045.409838] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.427531] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c185bd-8a93-bbe3-4f49-6ee770a89337, 'name': SearchDatastore_Task, 'duration_secs': 0.017799} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.428421] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60aa85fa-a1d9-4c59-9d31-6261cd7640fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.440090] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1045.440090] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527b0052-6b03-e2c7-aaef-44499f25b7bf" [ 1045.440090] env[62814]: _type = "Task" [ 1045.440090] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.446001] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527b0052-6b03-e2c7-aaef-44499f25b7bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.506196] env[62814]: DEBUG nova.network.neutron [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Successfully created port: 942dfa01-e165-41b2-8a4e-4d0d7dbb3f31 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1045.678534] env[62814]: DEBUG nova.network.neutron [req-c43ce6f7-dd82-49d8-b528-1c4db6fc7c1c req-6c24ec83-8906-4196-a005-eecbf52ce0df service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Updated VIF entry in instance network info cache for port 76bd40fc-ecff-43a9-a0e9-f58b9500f9c5. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1045.679025] env[62814]: DEBUG nova.network.neutron [req-c43ce6f7-dd82-49d8-b528-1c4db6fc7c1c req-6c24ec83-8906-4196-a005-eecbf52ce0df service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Updating instance_info_cache with network_info: [{"id": "76bd40fc-ecff-43a9-a0e9-f58b9500f9c5", "address": "fa:16:3e:f8:d9:df", "network": {"id": "c3a9b4ad-b825-451e-ad5a-4e0331bb2d16", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1261267397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e8761ca6d3444b2a94b18ef8628f4f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap76bd40fc-ec", "ovs_interfaceid": "76bd40fc-ecff-43a9-a0e9-f58b9500f9c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.703420] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1045.740436] env[62814]: DEBUG nova.network.neutron [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Successfully updated port: 89bae162-2360-4722-8b1c-ade210004fcc {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1045.747459] env[62814]: DEBUG nova.objects.instance [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lazy-loading 'flavor' on Instance uuid 316931ae-7a62-4bac-81e4-1fee9a36164e {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.859990] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294293, 'name': CreateVM_Task, 'duration_secs': 0.400793} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.860221] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1045.860939] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.861112] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1045.861423] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1045.861728] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c09756d-be10-4d5b-9aaf-4dbfbe077803 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.867468] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1045.867468] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52831c31-033f-7812-5f5e-e458ee903724" [ 1045.867468] env[62814]: _type = "Task" [ 1045.867468] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.876501] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52831c31-033f-7812-5f5e-e458ee903724, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.926262] env[62814]: INFO nova.compute.manager [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Took 38.57 seconds to build instance. [ 1045.947853] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527b0052-6b03-e2c7-aaef-44499f25b7bf, 'name': SearchDatastore_Task, 'duration_secs': 0.012603} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.948608] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1045.948608] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] cfab341e-57a1-48b4-9b6b-ceecf28c223b/cfab341e-57a1-48b4-9b6b-ceecf28c223b.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1045.949036] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e3ed9e54-4cb2-4747-ab4d-26067d30d269 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.957149] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1045.957149] env[62814]: value = "task-4294294" [ 1045.957149] env[62814]: _type = "Task" [ 1045.957149] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.965440] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294294, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.049786] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1046.050116] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1046.050851] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1046.050851] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1046.050851] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1046.050851] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1046.051173] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1046.051386] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1046.051651] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1046.051830] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1046.052021] env[62814]: DEBUG nova.virt.hardware [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1046.052960] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1aa7f8-485e-46bd-a054-ca8d19699f1f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.062611] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75bdf24-055a-4eb6-b4fc-e09176d91be0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.079657] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a3:d9:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '605f83bd-808c-4b54-922e-54b14690987a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '08e04ba7-4665-4982-ac59-6412cc953d93', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1046.087305] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1046.087624] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1046.087851] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dafe2f98-c762-4b07-832b-add1df9d0b02 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.107746] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1046.107746] env[62814]: value = "task-4294295" [ 1046.107746] env[62814]: _type = "Task" [ 1046.107746] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.116596] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294295, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.182062] env[62814]: DEBUG oslo_concurrency.lockutils [req-c43ce6f7-dd82-49d8-b528-1c4db6fc7c1c req-6c24ec83-8906-4196-a005-eecbf52ce0df service nova] Releasing lock "refresh_cache-1c86646f-6f38-4f8d-bea0-8a6b93680aba" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1046.209998] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1046.210272] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.027s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.210722] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 31.838s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.244549] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "refresh_cache-b4ee4cee-3298-4955-8375-8ca8c04b2f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.244789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired lock "refresh_cache-b4ee4cee-3298-4955-8375-8ca8c04b2f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1046.245017] env[62814]: DEBUG nova.network.neutron [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.252821] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c72076d-0556-43f4-9a94-9ecc2e8396ad tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 9.833s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.379863] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52831c31-033f-7812-5f5e-e458ee903724, 'name': SearchDatastore_Task, 'duration_secs': 0.012483} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.380307] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1046.381073] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1046.381073] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.381073] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1046.381449] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1046.381534] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1c39e61-d75c-46c3-a9b6-983d57761c4b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.400702] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1046.400914] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1046.401814] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6921c855-8faa-4f29-954e-01d22b35accd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.410078] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1046.410078] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520ebd2e-1f22-059e-f66a-a002f6d7f632" [ 1046.410078] env[62814]: _type = "Task" [ 1046.410078] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.416266] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "316931ae-7a62-4bac-81e4-1fee9a36164e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.416507] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.416714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "316931ae-7a62-4bac-81e4-1fee9a36164e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1046.416900] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1046.417084] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.422031] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520ebd2e-1f22-059e-f66a-a002f6d7f632, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.422594] env[62814]: INFO nova.compute.manager [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Terminating instance [ 1046.430286] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3149e8e6-7bff-45fc-b384-a83aa3853614 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lock "88ec5aba-f38d-4c75-af29-e3df3aa49640" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.086s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1046.471699] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294294, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.618777] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294295, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.716318] env[62814]: INFO nova.compute.claims [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1046.793885] env[62814]: DEBUG nova.network.neutron [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1046.920478] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520ebd2e-1f22-059e-f66a-a002f6d7f632, 'name': SearchDatastore_Task, 'duration_secs': 0.059916} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.921302] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5ac97ec-476d-4572-890d-9c73299c51a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.927712] env[62814]: DEBUG nova.compute.manager [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1046.927947] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1046.928269] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1046.928269] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209eaec-6625-28ca-9f27-caa0b2a97fb9" [ 1046.928269] env[62814]: _type = "Task" [ 1046.928269] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.928460] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5141115c-1ffe-411f-adfe-9dc31a46017e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.938730] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209eaec-6625-28ca-9f27-caa0b2a97fb9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.939958] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1046.939958] env[62814]: value = "task-4294296" [ 1046.939958] env[62814]: _type = "Task" [ 1046.939958] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.947530] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.949036] env[62814]: DEBUG nova.network.neutron [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Updating instance_info_cache with network_info: [{"id": "89bae162-2360-4722-8b1c-ade210004fcc", "address": "fa:16:3e:42:bc:4a", "network": {"id": "c3a9b4ad-b825-451e-ad5a-4e0331bb2d16", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1261267397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e8761ca6d3444b2a94b18ef8628f4f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89bae162-23", "ovs_interfaceid": "89bae162-2360-4722-8b1c-ade210004fcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.968366] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294294, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572789} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.968636] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] cfab341e-57a1-48b4-9b6b-ceecf28c223b/cfab341e-57a1-48b4-9b6b-ceecf28c223b.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1046.968849] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.969111] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b9d77be2-d878-4144-82d3-21dfe059a5d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.977127] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1046.977127] env[62814]: value = "task-4294297" [ 1046.977127] env[62814]: _type = "Task" [ 1046.977127] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.988285] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294297, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.123150] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294295, 'name': CreateVM_Task, 'duration_secs': 0.67039} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.123332] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1047.124012] env[62814]: DEBUG oslo_concurrency.lockutils [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.124180] env[62814]: DEBUG oslo_concurrency.lockutils [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.124489] env[62814]: DEBUG oslo_concurrency.lockutils [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1047.124744] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28182d7a-ea52-4138-8fc0-7d13181b6d78 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.129836] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1047.129836] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e0bb81-888f-efb9-babc-3d52180c4828" [ 1047.129836] env[62814]: _type = "Task" [ 1047.129836] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.137651] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e0bb81-888f-efb9-babc-3d52180c4828, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.175047] env[62814]: DEBUG nova.network.neutron [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Successfully updated port: 942dfa01-e165-41b2-8a4e-4d0d7dbb3f31 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1047.223205] env[62814]: INFO nova.compute.resource_tracker [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating resource usage from migration 953ceb03-e52d-4c0a-9b6d-70bce2963e8d [ 1047.422888] env[62814]: DEBUG nova.compute.manager [req-ae3b96a5-e099-4a66-94df-024460c078b7 req-f5cbf97a-564c-46db-ab72-6ad70fe5b2a2 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Received event network-vif-plugged-942dfa01-e165-41b2-8a4e-4d0d7dbb3f31 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1047.423198] env[62814]: DEBUG oslo_concurrency.lockutils [req-ae3b96a5-e099-4a66-94df-024460c078b7 req-f5cbf97a-564c-46db-ab72-6ad70fe5b2a2 service nova] Acquiring lock "885e0798-2fa6-4f6b-82be-517b6d1168d2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.423460] env[62814]: DEBUG oslo_concurrency.lockutils [req-ae3b96a5-e099-4a66-94df-024460c078b7 req-f5cbf97a-564c-46db-ab72-6ad70fe5b2a2 service nova] Lock "885e0798-2fa6-4f6b-82be-517b6d1168d2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.423719] env[62814]: DEBUG oslo_concurrency.lockutils [req-ae3b96a5-e099-4a66-94df-024460c078b7 req-f5cbf97a-564c-46db-ab72-6ad70fe5b2a2 service nova] Lock "885e0798-2fa6-4f6b-82be-517b6d1168d2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.423946] env[62814]: DEBUG nova.compute.manager [req-ae3b96a5-e099-4a66-94df-024460c078b7 req-f5cbf97a-564c-46db-ab72-6ad70fe5b2a2 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] No waiting events found dispatching network-vif-plugged-942dfa01-e165-41b2-8a4e-4d0d7dbb3f31 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1047.424184] env[62814]: WARNING nova.compute.manager [req-ae3b96a5-e099-4a66-94df-024460c078b7 req-f5cbf97a-564c-46db-ab72-6ad70fe5b2a2 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Received unexpected event network-vif-plugged-942dfa01-e165-41b2-8a4e-4d0d7dbb3f31 for instance with vm_state active and task_state None. [ 1047.441673] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209eaec-6625-28ca-9f27-caa0b2a97fb9, 'name': SearchDatastore_Task, 'duration_secs': 0.010552} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.447395] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.447748] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1c86646f-6f38-4f8d-bea0-8a6b93680aba/1c86646f-6f38-4f8d-bea0-8a6b93680aba.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1047.448959] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b20dbcf-14d4-431a-8958-61e6dc26b91e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.451687] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Releasing lock "refresh_cache-b4ee4cee-3298-4955-8375-8ca8c04b2f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.451935] env[62814]: DEBUG nova.compute.manager [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Instance network_info: |[{"id": "89bae162-2360-4722-8b1c-ade210004fcc", "address": "fa:16:3e:42:bc:4a", "network": {"id": "c3a9b4ad-b825-451e-ad5a-4e0331bb2d16", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1261267397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e8761ca6d3444b2a94b18ef8628f4f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89bae162-23", "ovs_interfaceid": "89bae162-2360-4722-8b1c-ade210004fcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1047.452334] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:42:bc:4a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '89bae162-2360-4722-8b1c-ade210004fcc', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1047.460090] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1047.466279] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1047.466901] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.468422] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b66cb8f2-2bb2-4714-b533-6dbe2eb7eccf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.484717] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1047.484717] env[62814]: value = "task-4294298" [ 1047.484717] env[62814]: _type = "Task" [ 1047.484717] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.497610] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1047.497610] env[62814]: value = "task-4294299" [ 1047.497610] env[62814]: _type = "Task" [ 1047.497610] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.501771] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294298, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.511956] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294297, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.518559] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294299, 'name': CreateVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.572809] env[62814]: DEBUG nova.compute.manager [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Received event network-vif-plugged-89bae162-2360-4722-8b1c-ade210004fcc {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1047.572989] env[62814]: DEBUG oslo_concurrency.lockutils [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] Acquiring lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1047.573233] env[62814]: DEBUG oslo_concurrency.lockutils [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] Lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1047.573406] env[62814]: DEBUG oslo_concurrency.lockutils [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] Lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1047.573577] env[62814]: DEBUG nova.compute.manager [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] No waiting events found dispatching network-vif-plugged-89bae162-2360-4722-8b1c-ade210004fcc {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1047.573745] env[62814]: WARNING nova.compute.manager [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Received unexpected event network-vif-plugged-89bae162-2360-4722-8b1c-ade210004fcc for instance with vm_state building and task_state spawning. [ 1047.573931] env[62814]: DEBUG nova.compute.manager [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Received event network-changed-89bae162-2360-4722-8b1c-ade210004fcc {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1047.574115] env[62814]: DEBUG nova.compute.manager [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Refreshing instance network info cache due to event network-changed-89bae162-2360-4722-8b1c-ade210004fcc. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1047.574303] env[62814]: DEBUG oslo_concurrency.lockutils [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] Acquiring lock "refresh_cache-b4ee4cee-3298-4955-8375-8ca8c04b2f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.574438] env[62814]: DEBUG oslo_concurrency.lockutils [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] Acquired lock "refresh_cache-b4ee4cee-3298-4955-8375-8ca8c04b2f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.574603] env[62814]: DEBUG nova.network.neutron [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Refreshing network info cache for port 89bae162-2360-4722-8b1c-ade210004fcc {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1047.606041] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a86cf619-8096-45f0-a8e7-fb3f05791729 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.615015] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74e2df9-fdf8-4ac7-936e-a6b89931ed5c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.651386] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc8255fc-c3de-4467-831c-5091caf3c038 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.663731] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2fa7cf-d8f9-4468-9e19-9816a44b4bd2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.667539] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e0bb81-888f-efb9-babc-3d52180c4828, 'name': SearchDatastore_Task, 'duration_secs': 0.010713} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.667906] env[62814]: DEBUG oslo_concurrency.lockutils [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1047.668176] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1047.668436] env[62814]: DEBUG oslo_concurrency.lockutils [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.668581] env[62814]: DEBUG oslo_concurrency.lockutils [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.668757] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1047.669472] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3ffd6d1-41b0-448d-b121-8d31183ec6fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.682053] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.682053] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1047.682053] env[62814]: DEBUG nova.network.neutron [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1047.684151] env[62814]: DEBUG nova.compute.provider_tree [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.690556] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1047.690731] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1047.691557] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2397d29f-ffae-4e42-ae93-f43e4fa6ce90 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.699170] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1047.699170] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e16500-ce9d-1bf7-3a0d-88127d503b12" [ 1047.699170] env[62814]: _type = "Task" [ 1047.699170] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.709238] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e16500-ce9d-1bf7-3a0d-88127d503b12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.955399] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294296, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.006303] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294298, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.006575] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294297, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.966275} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.010736] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1048.011680] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82ead90-dd59-44f0-847c-5c117d37dde6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.020305] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294299, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.040766] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] cfab341e-57a1-48b4-9b6b-ceecf28c223b/cfab341e-57a1-48b4-9b6b-ceecf28c223b.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1048.041103] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2bfb9085-09a7-4803-87c9-d564f8dfee4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.063225] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1048.063225] env[62814]: value = "task-4294300" [ 1048.063225] env[62814]: _type = "Task" [ 1048.063225] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.073321] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294300, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.190085] env[62814]: DEBUG nova.scheduler.client.report [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1048.212581] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e16500-ce9d-1bf7-3a0d-88127d503b12, 'name': SearchDatastore_Task, 'duration_secs': 0.064836} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.213747] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62fcd5ef-334f-49b3-8f6c-369f3572c81d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.223329] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1048.223329] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524aff51-b6fd-e374-112e-a2cce78c4374" [ 1048.223329] env[62814]: _type = "Task" [ 1048.223329] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.233108] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524aff51-b6fd-e374-112e-a2cce78c4374, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.248250] env[62814]: WARNING nova.network.neutron [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] 1de03cb0-69fa-4200-8fa1-26375f44369a already exists in list: networks containing: ['1de03cb0-69fa-4200-8fa1-26375f44369a']. ignoring it [ 1048.365035] env[62814]: DEBUG nova.network.neutron [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Updated VIF entry in instance network info cache for port 89bae162-2360-4722-8b1c-ade210004fcc. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.365406] env[62814]: DEBUG nova.network.neutron [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Updating instance_info_cache with network_info: [{"id": "89bae162-2360-4722-8b1c-ade210004fcc", "address": "fa:16:3e:42:bc:4a", "network": {"id": "c3a9b4ad-b825-451e-ad5a-4e0331bb2d16", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1261267397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e8761ca6d3444b2a94b18ef8628f4f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap89bae162-23", "ovs_interfaceid": "89bae162-2360-4722-8b1c-ade210004fcc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.435938] env[62814]: DEBUG oslo_concurrency.lockutils [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquiring lock "88ec5aba-f38d-4c75-af29-e3df3aa49640" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.436327] env[62814]: DEBUG oslo_concurrency.lockutils [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lock "88ec5aba-f38d-4c75-af29-e3df3aa49640" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.436459] env[62814]: DEBUG oslo_concurrency.lockutils [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquiring lock "88ec5aba-f38d-4c75-af29-e3df3aa49640-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1048.436660] env[62814]: DEBUG oslo_concurrency.lockutils [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lock "88ec5aba-f38d-4c75-af29-e3df3aa49640-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.436869] env[62814]: DEBUG oslo_concurrency.lockutils [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lock "88ec5aba-f38d-4c75-af29-e3df3aa49640-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.439249] env[62814]: INFO nova.compute.manager [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Terminating instance [ 1048.453434] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294296, 'name': PowerOffVM_Task, 'duration_secs': 1.23843} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.453693] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1048.453920] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1048.454145] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845794', 'volume_id': '4878e7fc-22eb-4976-bdda-753fce367d65', 'name': 'volume-4878e7fc-22eb-4976-bdda-753fce367d65', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '316931ae-7a62-4bac-81e4-1fee9a36164e', 'attached_at': '', 'detached_at': '', 'volume_id': '4878e7fc-22eb-4976-bdda-753fce367d65', 'serial': '4878e7fc-22eb-4976-bdda-753fce367d65'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1048.454977] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d13b5c65-329c-41fb-8971-287e6830a79a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.478865] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62626f79-80f0-4958-a214-a9933c2a03ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.488655] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-846d4d33-1085-46df-bbae-f38aa4e0594a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.522401] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294298, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556523} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.523213] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1c86646f-6f38-4f8d-bea0-8a6b93680aba/1c86646f-6f38-4f8d-bea0-8a6b93680aba.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1048.523339] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1048.527438] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fb3ca9-a89c-4151-8310-724b92f17989 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.531990] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-41b26f11-b0e2-4212-89ba-317a0788ba85 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.540790] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294299, 'name': CreateVM_Task, 'duration_secs': 0.762056} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.552657] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1048.553334] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1048.553334] env[62814]: value = "task-4294301" [ 1048.553334] env[62814]: _type = "Task" [ 1048.553334] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.553534] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] The volume has not been displaced from its original location: [datastore1] volume-4878e7fc-22eb-4976-bdda-753fce367d65/volume-4878e7fc-22eb-4976-bdda-753fce367d65.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1048.559347] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Reconfiguring VM instance instance-00000052 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1048.562967] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.563224] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.563555] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1048.564228] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f17c845a-b2e9-4bc7-be61-548bb22d061d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.577783] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f72636c-f833-4290-823d-f30f36e8f7c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.592539] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294300, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.592881] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1048.592881] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524311d5-5aeb-8e1f-ab9b-986959025684" [ 1048.592881] env[62814]: _type = "Task" [ 1048.592881] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.596016] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294301, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.600371] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1048.600371] env[62814]: value = "task-4294302" [ 1048.600371] env[62814]: _type = "Task" [ 1048.600371] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.607588] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524311d5-5aeb-8e1f-ab9b-986959025684, 'name': SearchDatastore_Task, 'duration_secs': 0.011245} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.608278] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1048.608518] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1048.608733] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.614255] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294302, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.698506] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.488s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.698712] env[62814]: INFO nova.compute.manager [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Migrating [ 1048.708016] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.887s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.708332] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.710196] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.675s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.710308] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.712043] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.905s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.712235] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1048.714191] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.172s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1048.715690] env[62814]: INFO nova.compute.claims [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1048.738288] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524aff51-b6fd-e374-112e-a2cce78c4374, 'name': SearchDatastore_Task, 'duration_secs': 0.010898} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.741681] env[62814]: DEBUG oslo_concurrency.lockutils [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1048.742012] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a07b964f-fe88-4c88-b6f4-61ed0973716c/a07b964f-fe88-4c88-b6f4-61ed0973716c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1048.742750] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1048.742957] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1048.743388] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea8099d4-c979-4596-acfb-4ec05fd6c4a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.746190] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5629ff17-e7c1-4224-a54e-077e33f144e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.753386] env[62814]: INFO nova.scheduler.client.report [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleted allocations for instance 40e2d845-0211-4c84-aef7-94014f999e1d [ 1048.758832] env[62814]: INFO nova.scheduler.client.report [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Deleted allocations for instance d9242042-6209-4b04-bf00-00dd04d9d6a0 [ 1048.761524] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1048.761524] env[62814]: value = "task-4294303" [ 1048.761524] env[62814]: _type = "Task" [ 1048.761524] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.765113] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1048.765316] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1048.770470] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce1019b6-8fdb-4669-afeb-bd189e5c0235 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.773808] env[62814]: INFO nova.scheduler.client.report [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Deleted allocations for instance fd1867e3-2523-4969-a157-b14c650f3779 [ 1048.782083] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294303, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.784297] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1048.784297] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527087e6-23c8-f21c-1635-358653668dfc" [ 1048.784297] env[62814]: _type = "Task" [ 1048.784297] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.796674] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527087e6-23c8-f21c-1635-358653668dfc, 'name': SearchDatastore_Task, 'duration_secs': 0.012359} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.800167] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29716ab2-0862-455a-8a01-eba61b5c2ff7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.809332] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1048.809332] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526998a1-18eb-a8c1-8865-e3d4bf353597" [ 1048.809332] env[62814]: _type = "Task" [ 1048.809332] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.819169] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526998a1-18eb-a8c1-8865-e3d4bf353597, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.870361] env[62814]: DEBUG oslo_concurrency.lockutils [req-e10d62f0-5c81-40eb-96ba-0c9a065ef4a9 req-8e2d6e33-1242-4eed-90bf-9aa7404d3941 service nova] Releasing lock "refresh_cache-b4ee4cee-3298-4955-8375-8ca8c04b2f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1048.920650] env[62814]: DEBUG nova.network.neutron [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updating instance_info_cache with network_info: [{"id": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "address": "fa:16:3e:ee:a8:e0", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b18381-f5", "ovs_interfaceid": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "942dfa01-e165-41b2-8a4e-4d0d7dbb3f31", "address": "fa:16:3e:2e:33:ff", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap942dfa01-e1", "ovs_interfaceid": "942dfa01-e165-41b2-8a4e-4d0d7dbb3f31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.947941] env[62814]: DEBUG nova.compute.manager [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1048.948246] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1048.949504] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb548c2-db9a-4201-b352-803400a4da9f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.960448] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1048.960789] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24188c73-0e04-4964-bca8-138820bb1735 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.969638] env[62814]: DEBUG oslo_vmware.api [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for the task: (returnval){ [ 1048.969638] env[62814]: value = "task-4294304" [ 1048.969638] env[62814]: _type = "Task" [ 1048.969638] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.982996] env[62814]: DEBUG oslo_vmware.api [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294304, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.077818] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294301, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.20523} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.077818] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1049.080024] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d71944-c0bf-4085-b3ff-966181b7e34e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.094039] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294300, 'name': ReconfigVM_Task, 'duration_secs': 0.636731} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.104430] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Reconfigured VM instance instance-00000055 to attach disk [datastore2] cfab341e-57a1-48b4-9b6b-ceecf28c223b/cfab341e-57a1-48b4-9b6b-ceecf28c223b.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.115883] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 1c86646f-6f38-4f8d-bea0-8a6b93680aba/1c86646f-6f38-4f8d-bea0-8a6b93680aba.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.116324] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-47599add-d44c-489e-a51b-68bd7ee7f4d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.121571] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0c50dcb-b8cc-4248-bd47-e904f92e61c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.146928] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294302, 'name': ReconfigVM_Task, 'duration_secs': 0.405156} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.152025] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Reconfigured VM instance instance-00000052 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1049.156807] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1049.156807] env[62814]: value = "task-4294305" [ 1049.156807] env[62814]: _type = "Task" [ 1049.156807] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.157208] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1049.157208] env[62814]: value = "task-4294306" [ 1049.157208] env[62814]: _type = "Task" [ 1049.157208] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.157499] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-572ce298-ba93-4a83-ae4f-5c1de85a30fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.188553] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294305, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.189823] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1049.189823] env[62814]: value = "task-4294307" [ 1049.189823] env[62814]: _type = "Task" [ 1049.189823] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.192976] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294306, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.206591] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294307, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.227047] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.227372] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1049.227696] env[62814]: DEBUG nova.network.neutron [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1049.265957] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e8e75611-45bf-43b8-a382-1de30cf6ab27 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "40e2d845-0211-4c84-aef7-94014f999e1d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.769s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.274642] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ea3f13f-f6af-4ec2-a279-025f8ad54558 tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "d9242042-6209-4b04-bf00-00dd04d9d6a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.097s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.281423] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294303, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.286121] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7df9a199-9a31-4692-a4fe-bacba55d0ca7 tempest-ImagesTestJSON-385848730 tempest-ImagesTestJSON-385848730-project-member] Lock "fd1867e3-2523-4969-a157-b14c650f3779" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 37.071s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1049.321491] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526998a1-18eb-a8c1-8865-e3d4bf353597, 'name': SearchDatastore_Task, 'duration_secs': 0.015321} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.321943] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.322206] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b4ee4cee-3298-4955-8375-8ca8c04b2f9f/b4ee4cee-3298-4955-8375-8ca8c04b2f9f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1049.322488] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec25b6e3-25f2-4d4e-86ee-da9c54845d16 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.330178] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1049.330178] env[62814]: value = "task-4294308" [ 1049.330178] env[62814]: _type = "Task" [ 1049.330178] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.340019] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.427723] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1049.428274] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.428379] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1049.429229] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff9b43e-f83d-405e-ad5f-e009432531a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.449050] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1049.449458] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1049.449751] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1049.450074] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1049.450313] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1049.450547] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1049.451154] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1049.451266] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1049.451501] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1049.451720] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1049.451948] env[62814]: DEBUG nova.virt.hardware [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1049.458886] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Reconfiguring VM to attach interface {{(pid=62814) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1049.459275] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7b572fb-aa55-428c-aba5-f25c8bba3882 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.482376] env[62814]: DEBUG oslo_vmware.api [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294304, 'name': PowerOffVM_Task, 'duration_secs': 0.35574} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.483849] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1049.484469] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1049.484469] env[62814]: DEBUG oslo_vmware.api [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1049.484469] env[62814]: value = "task-4294309" [ 1049.484469] env[62814]: _type = "Task" [ 1049.484469] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.484726] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6976bdc3-3f3b-4fab-a457-51fdb9a824d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.496479] env[62814]: DEBUG oslo_vmware.api [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294309, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.614710] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.614970] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.615220] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Deleting the datastore file [datastore2] 88ec5aba-f38d-4c75-af29-e3df3aa49640 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.615813] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd1e7213-d5a1-489a-9122-d24e5f91fb02 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.625641] env[62814]: DEBUG oslo_vmware.api [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for the task: (returnval){ [ 1049.625641] env[62814]: value = "task-4294311" [ 1049.625641] env[62814]: _type = "Task" [ 1049.625641] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.638477] env[62814]: DEBUG oslo_vmware.api [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294311, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.685228] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294306, 'name': ReconfigVM_Task, 'duration_secs': 0.402139} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.688543] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 1c86646f-6f38-4f8d-bea0-8a6b93680aba/1c86646f-6f38-4f8d-bea0-8a6b93680aba.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1049.689419] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294305, 'name': Rename_Task, 'duration_secs': 0.251862} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.689714] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2908df4d-3fc6-460d-81e6-05a173faf5c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.691757] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1049.692043] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4bb56094-e0bd-4862-b284-761a31f683ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.705156] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1049.705156] env[62814]: value = "task-4294312" [ 1049.705156] env[62814]: _type = "Task" [ 1049.705156] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.708080] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294307, 'name': ReconfigVM_Task, 'duration_secs': 0.212298} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.708559] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1049.708559] env[62814]: value = "task-4294313" [ 1049.708559] env[62814]: _type = "Task" [ 1049.708559] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.712117] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845794', 'volume_id': '4878e7fc-22eb-4976-bdda-753fce367d65', 'name': 'volume-4878e7fc-22eb-4976-bdda-753fce367d65', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '316931ae-7a62-4bac-81e4-1fee9a36164e', 'attached_at': '', 'detached_at': '', 'volume_id': '4878e7fc-22eb-4976-bdda-753fce367d65', 'serial': '4878e7fc-22eb-4976-bdda-753fce367d65'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1049.712512] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1049.713426] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51e0efd-e5ef-4949-8463-027f716bec29 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.723851] env[62814]: DEBUG nova.compute.manager [req-796caa9b-1ebd-498d-b007-931042b9f4b5 req-acd232e2-2bd8-4da5-93ab-2ee3342cb996 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Received event network-changed-942dfa01-e165-41b2-8a4e-4d0d7dbb3f31 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1049.724190] env[62814]: DEBUG nova.compute.manager [req-796caa9b-1ebd-498d-b007-931042b9f4b5 req-acd232e2-2bd8-4da5-93ab-2ee3342cb996 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Refreshing instance network info cache due to event network-changed-942dfa01-e165-41b2-8a4e-4d0d7dbb3f31. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1049.724475] env[62814]: DEBUG oslo_concurrency.lockutils [req-796caa9b-1ebd-498d-b007-931042b9f4b5 req-acd232e2-2bd8-4da5-93ab-2ee3342cb996 service nova] Acquiring lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.724639] env[62814]: DEBUG oslo_concurrency.lockutils [req-796caa9b-1ebd-498d-b007-931042b9f4b5 req-acd232e2-2bd8-4da5-93ab-2ee3342cb996 service nova] Acquired lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1049.724834] env[62814]: DEBUG nova.network.neutron [req-796caa9b-1ebd-498d-b007-931042b9f4b5 req-acd232e2-2bd8-4da5-93ab-2ee3342cb996 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Refreshing network info cache for port 942dfa01-e165-41b2-8a4e-4d0d7dbb3f31 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1049.732847] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294312, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.748424] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294313, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.748729] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1049.750298] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cda9a76e-b624-4220-a183-0b812eefefa3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.787266] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294303, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546376} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.787585] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] a07b964f-fe88-4c88-b6f4-61ed0973716c/a07b964f-fe88-4c88-b6f4-61ed0973716c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1049.787879] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.788508] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7940a1f3-3720-4a83-a04c-bc594099f0d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.800319] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1049.800319] env[62814]: value = "task-4294315" [ 1049.800319] env[62814]: _type = "Task" [ 1049.800319] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.818211] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294315, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.845017] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1049.845299] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1049.845450] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleting the datastore file [datastore2] 316931ae-7a62-4bac-81e4-1fee9a36164e {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1049.850742] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e8e887b-0165-416d-a23f-04f56fcd5b7f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.853071] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294308, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49746} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.855829] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b4ee4cee-3298-4955-8375-8ca8c04b2f9f/b4ee4cee-3298-4955-8375-8ca8c04b2f9f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1049.856087] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.858281] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44301ca3-fc6c-40d2-aa31-e01ca7d63a42 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.865062] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1049.865062] env[62814]: value = "task-4294316" [ 1049.865062] env[62814]: _type = "Task" [ 1049.865062] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.872325] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1049.872325] env[62814]: value = "task-4294317" [ 1049.872325] env[62814]: _type = "Task" [ 1049.872325] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.886705] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294316, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.897735] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294317, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.998098] env[62814]: DEBUG oslo_vmware.api [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294309, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.140292] env[62814]: DEBUG oslo_vmware.api [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Task: {'id': task-4294311, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.357224} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.141457] env[62814]: DEBUG nova.network.neutron [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance_info_cache with network_info: [{"id": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "address": "fa:16:3e:93:37:3e", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa00402c1-eb", "ovs_interfaceid": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.143554] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.143768] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.143969] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.144318] env[62814]: INFO nova.compute.manager [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1050.144575] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1050.147315] env[62814]: DEBUG nova.compute.manager [-] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1050.147432] env[62814]: DEBUG nova.network.neutron [-] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1050.219881] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ebe5ba-b2e4-4863-9ac2-7620d05bb7fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.231851] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294313, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.238883] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294312, 'name': Rename_Task, 'duration_secs': 0.257381} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.239283] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1050.240339] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9c608b-cbc5-47fd-8a94-d20a00c8203f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.244666] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ccd6d9f-87f8-48e1-bdda-dc0fce33ee8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.280386] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30820fb4-29b8-4656-9c9f-5aea0986d731 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.283565] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1050.283565] env[62814]: value = "task-4294318" [ 1050.283565] env[62814]: _type = "Task" [ 1050.283565] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.303047] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e868a06-1631-4823-b38f-612c2e032dbd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.307568] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294318, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.323619] env[62814]: DEBUG nova.compute.provider_tree [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1050.329085] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294315, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084419} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.329396] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.330301] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded9ad38-abeb-474c-a954-54e725ef7fda {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.359427] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] a07b964f-fe88-4c88-b6f4-61ed0973716c/a07b964f-fe88-4c88-b6f4-61ed0973716c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.362597] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-094b00a3-ed8e-45d6-9158-b7064907291a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.399535] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294317, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08061} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.399535] env[62814]: DEBUG oslo_vmware.api [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294316, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.167307} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.401031] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.401031] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.401175] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.401348] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.401519] env[62814]: INFO nova.compute.manager [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Took 3.47 seconds to destroy the instance on the hypervisor. [ 1050.401786] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1050.402045] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1050.402045] env[62814]: value = "task-4294319" [ 1050.402045] env[62814]: _type = "Task" [ 1050.402045] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.402963] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71179dec-7456-4df1-8207-76dea26e8c7d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.405670] env[62814]: DEBUG nova.compute.manager [-] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1050.405670] env[62814]: DEBUG nova.network.neutron [-] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1050.421370] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294319, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.442582] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] b4ee4cee-3298-4955-8375-8ca8c04b2f9f/b4ee4cee-3298-4955-8375-8ca8c04b2f9f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.443381] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e357f4c-f913-41b6-9808-47a9092d734c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.470870] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1050.470870] env[62814]: value = "task-4294320" [ 1050.470870] env[62814]: _type = "Task" [ 1050.470870] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.480331] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294320, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.498063] env[62814]: DEBUG oslo_vmware.api [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294309, 'name': ReconfigVM_Task, 'duration_secs': 0.830728} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.498752] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1050.499069] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Reconfigured VM to attach interface {{(pid=62814) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1050.592054] env[62814]: DEBUG nova.network.neutron [req-796caa9b-1ebd-498d-b007-931042b9f4b5 req-acd232e2-2bd8-4da5-93ab-2ee3342cb996 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updated VIF entry in instance network info cache for port 942dfa01-e165-41b2-8a4e-4d0d7dbb3f31. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1050.592498] env[62814]: DEBUG nova.network.neutron [req-796caa9b-1ebd-498d-b007-931042b9f4b5 req-acd232e2-2bd8-4da5-93ab-2ee3342cb996 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updating instance_info_cache with network_info: [{"id": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "address": "fa:16:3e:ee:a8:e0", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b18381-f5", "ovs_interfaceid": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "942dfa01-e165-41b2-8a4e-4d0d7dbb3f31", "address": "fa:16:3e:2e:33:ff", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap942dfa01-e1", "ovs_interfaceid": "942dfa01-e165-41b2-8a4e-4d0d7dbb3f31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.644187] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1050.733762] env[62814]: DEBUG oslo_vmware.api [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294313, 'name': PowerOnVM_Task, 'duration_secs': 0.710606} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.734702] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1050.735088] env[62814]: INFO nova.compute.manager [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Took 12.68 seconds to spawn the instance on the hypervisor. [ 1050.735250] env[62814]: DEBUG nova.compute.manager [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1050.736597] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2473bea8-d383-40c1-b55a-4ddecff3028c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.798836] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294318, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.849242] env[62814]: ERROR nova.scheduler.client.report [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [req-9d7d625c-809e-43c4-bb54-7ba2e9c8e237] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9d7d625c-809e-43c4-bb54-7ba2e9c8e237"}]} [ 1050.867436] env[62814]: DEBUG nova.scheduler.client.report [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1050.884869] env[62814]: DEBUG nova.scheduler.client.report [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1050.885420] env[62814]: DEBUG nova.compute.provider_tree [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1050.901049] env[62814]: DEBUG nova.scheduler.client.report [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1050.913472] env[62814]: DEBUG nova.network.neutron [-] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1050.923150] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294319, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.924495] env[62814]: DEBUG nova.scheduler.client.report [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1050.985364] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294320, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.007728] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8b6db2bb-b5f5-41e7-ad35-811b88d0bff3 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-885e0798-2fa6-4f6b-82be-517b6d1168d2-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.022s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1051.095795] env[62814]: DEBUG oslo_concurrency.lockutils [req-796caa9b-1ebd-498d-b007-931042b9f4b5 req-acd232e2-2bd8-4da5-93ab-2ee3342cb996 service nova] Releasing lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1051.229373] env[62814]: DEBUG nova.network.neutron [-] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1051.249958] env[62814]: DEBUG nova.compute.manager [req-f7449011-d3a9-4686-be49-0a8e4e719100 req-cc6e0ca0-a448-4f90-b473-5b6d5b993dcd service nova] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Received event network-vif-deleted-031fa2fc-70be-4d13-bb5d-8c516281f35e {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1051.257729] env[62814]: INFO nova.compute.manager [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Took 39.36 seconds to build instance. [ 1051.271545] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c49a4f-cd89-4d86-9bea-1b08df74773c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.282637] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0b2685-48d9-4e72-b920-7606e3428f3d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.317551] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba8a44d-41be-4734-9488-a9afe1dd66c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.323762] env[62814]: DEBUG oslo_vmware.api [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294318, 'name': PowerOnVM_Task, 'duration_secs': 0.640577} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.324410] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1051.324653] env[62814]: INFO nova.compute.manager [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Took 10.64 seconds to spawn the instance on the hypervisor. [ 1051.324853] env[62814]: DEBUG nova.compute.manager [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1051.325620] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df08d6ae-4ffc-4315-8607-88d8c713aedd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.331924] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf702c7d-cec6-4cd1-b03d-cf5799871f39 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.350026] env[62814]: DEBUG nova.compute.provider_tree [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1051.418054] env[62814]: INFO nova.compute.manager [-] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Took 1.27 seconds to deallocate network for instance. [ 1051.418386] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294319, 'name': ReconfigVM_Task, 'duration_secs': 0.716344} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.420442] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Reconfigured VM instance instance-00000053 to attach disk [datastore2] a07b964f-fe88-4c88-b6f4-61ed0973716c/a07b964f-fe88-4c88-b6f4-61ed0973716c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.420848] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ec2aa0e-d0bf-4e19-938a-e042ad5ac64c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.432348] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1051.432348] env[62814]: value = "task-4294321" [ 1051.432348] env[62814]: _type = "Task" [ 1051.432348] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.441441] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294321, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.482251] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294320, 'name': ReconfigVM_Task, 'duration_secs': 0.785921} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.482533] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Reconfigured VM instance instance-00000057 to attach disk [datastore2] b4ee4cee-3298-4955-8375-8ca8c04b2f9f/b4ee4cee-3298-4955-8375-8ca8c04b2f9f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.483207] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79e2377b-1e1c-4f95-b7a3-2b079825c1b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.491053] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1051.491053] env[62814]: value = "task-4294322" [ 1051.491053] env[62814]: _type = "Task" [ 1051.491053] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.501826] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "1e259ec6-d31a-453b-87e0-baa446665d56" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1051.502113] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1051.503502] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294322, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.617883] env[62814]: DEBUG oslo_concurrency.lockutils [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "c7cef7f3-11db-44e1-a454-98830b465b52" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1051.618223] env[62814]: DEBUG oslo_concurrency.lockutils [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "c7cef7f3-11db-44e1-a454-98830b465b52" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1051.618437] env[62814]: DEBUG oslo_concurrency.lockutils [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "c7cef7f3-11db-44e1-a454-98830b465b52-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1051.618621] env[62814]: DEBUG oslo_concurrency.lockutils [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "c7cef7f3-11db-44e1-a454-98830b465b52-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1051.618790] env[62814]: DEBUG oslo_concurrency.lockutils [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "c7cef7f3-11db-44e1-a454-98830b465b52-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1051.621057] env[62814]: INFO nova.compute.manager [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Terminating instance [ 1051.736516] env[62814]: INFO nova.compute.manager [-] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Took 1.33 seconds to deallocate network for instance. [ 1051.762785] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1141d2e6-3780-4fed-9b53-f5222de7bb4d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.876s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1051.847097] env[62814]: INFO nova.compute.manager [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Took 39.35 seconds to build instance. [ 1051.885629] env[62814]: DEBUG nova.scheduler.client.report [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 131 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1051.885916] env[62814]: DEBUG nova.compute.provider_tree [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 131 to 132 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1051.886118] env[62814]: DEBUG nova.compute.provider_tree [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1051.928114] env[62814]: DEBUG oslo_concurrency.lockutils [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1051.944637] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294321, 'name': Rename_Task, 'duration_secs': 0.18036} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.944964] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1051.945241] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4658f0ee-8de4-41b2-8b22-49d50bcdff61 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.954016] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1051.954016] env[62814]: value = "task-4294323" [ 1051.954016] env[62814]: _type = "Task" [ 1051.954016] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.962230] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294323, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.002725] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294322, 'name': Rename_Task, 'duration_secs': 0.172703} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.003083] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1052.003349] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-713ef322-abbd-45dc-ae0e-1830e16a2a8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.005497] env[62814]: INFO nova.compute.manager [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Detaching volume 4bf78821-2028-4edc-a038-062cb9d09c71 [ 1052.015510] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1052.015510] env[62814]: value = "task-4294324" [ 1052.015510] env[62814]: _type = "Task" [ 1052.015510] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.023833] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294324, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.040737] env[62814]: INFO nova.virt.block_device [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Attempting to driver detach volume 4bf78821-2028-4edc-a038-062cb9d09c71 from mountpoint /dev/sdb [ 1052.041018] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1052.041221] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845740', 'volume_id': '4bf78821-2028-4edc-a038-062cb9d09c71', 'name': 'volume-4bf78821-2028-4edc-a038-062cb9d09c71', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1e259ec6-d31a-453b-87e0-baa446665d56', 'attached_at': '', 'detached_at': '', 'volume_id': '4bf78821-2028-4edc-a038-062cb9d09c71', 'serial': '4bf78821-2028-4edc-a038-062cb9d09c71'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1052.042594] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2162c522-2cb2-4dd1-8fff-0b860411dc11 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.067529] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5912ceac-f217-42db-9015-6508f2b91110 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.075549] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c16cc712-4031-4fce-999c-af9bf2eb1542 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.097877] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee4ca67-41f4-46b4-820e-aa3e15698098 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.117672] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] The volume has not been displaced from its original location: [datastore1] volume-4bf78821-2028-4edc-a038-062cb9d09c71/volume-4bf78821-2028-4edc-a038-062cb9d09c71.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1052.123204] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Reconfiguring VM instance instance-0000003c to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1052.124101] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cdfa8ba-252b-4232-967f-a32845c02233 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.137030] env[62814]: DEBUG nova.compute.manager [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1052.137249] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1052.138501] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4cf1ee-619f-4d30-ae48-6700fcef8519 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.147922] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1052.149305] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48de1882-faae-4b4d-a792-1d9c570b8c30 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.151099] env[62814]: DEBUG oslo_vmware.api [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1052.151099] env[62814]: value = "task-4294325" [ 1052.151099] env[62814]: _type = "Task" [ 1052.151099] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.159975] env[62814]: DEBUG oslo_vmware.api [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 1052.159975] env[62814]: value = "task-4294326" [ 1052.159975] env[62814]: _type = "Task" [ 1052.159975] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.169227] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067fe182-c96f-44c3-adc8-cd33c3b321dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.180075] env[62814]: DEBUG oslo_vmware.api [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294326, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.194879] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance '59d21ef7-df97-49ac-9329-4c18df6dd087' progress to 0 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1052.283498] env[62814]: INFO nova.compute.manager [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Took 0.55 seconds to detach 1 volumes for instance. [ 1052.348983] env[62814]: DEBUG oslo_concurrency.lockutils [None req-01d86e00-c87f-4334-bbd0-184234562b9e tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.871s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.391883] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.678s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.392451] env[62814]: DEBUG nova.compute.manager [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1052.398238] env[62814]: DEBUG oslo_concurrency.lockutils [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 32.707s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.398238] env[62814]: DEBUG oslo_concurrency.lockutils [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.401891] env[62814]: DEBUG oslo_concurrency.lockutils [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.657s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.402090] env[62814]: DEBUG oslo_concurrency.lockutils [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.404647] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.597s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1052.407048] env[62814]: INFO nova.compute.claims [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.437522] env[62814]: INFO nova.scheduler.client.report [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Deleted allocations for instance 508321ab-ce10-4953-a9e3-193b9975bec7 [ 1052.450127] env[62814]: INFO nova.scheduler.client.report [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Deleted allocations for instance a68901a9-9a9b-4127-bca9-64f98dfb151f [ 1052.475074] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294323, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.533115] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294324, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.661250] env[62814]: DEBUG oslo_vmware.api [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294325, 'name': ReconfigVM_Task, 'duration_secs': 0.299419} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.661541] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Reconfigured VM instance instance-0000003c to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1052.666693] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e70bf2e-a148-48eb-839f-e9711f9ef95f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.687701] env[62814]: DEBUG oslo_vmware.api [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294326, 'name': PowerOffVM_Task, 'duration_secs': 0.180004} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.689282] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1052.689468] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1052.689807] env[62814]: DEBUG oslo_vmware.api [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1052.689807] env[62814]: value = "task-4294327" [ 1052.689807] env[62814]: _type = "Task" [ 1052.689807] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.690112] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05ca26f7-5dfc-4846-8aeb-b9d5ab800540 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.701922] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1052.702209] env[62814]: DEBUG oslo_vmware.api [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294327, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.702428] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9c39fbe-5157-4e2f-a082-a773eb474b67 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.713456] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1052.713456] env[62814]: value = "task-4294329" [ 1052.713456] env[62814]: _type = "Task" [ 1052.713456] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.723896] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294329, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.772931] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1052.773188] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1052.773363] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Deleting the datastore file [datastore2] c7cef7f3-11db-44e1-a454-98830b465b52 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.773658] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d4efc02f-7fcb-462b-9add-ce7f5a39464a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.783682] env[62814]: DEBUG oslo_vmware.api [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for the task: (returnval){ [ 1052.783682] env[62814]: value = "task-4294330" [ 1052.783682] env[62814]: _type = "Task" [ 1052.783682] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.792413] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1052.798738] env[62814]: DEBUG oslo_vmware.api [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294330, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.913647] env[62814]: DEBUG nova.compute.utils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1052.919713] env[62814]: DEBUG nova.compute.manager [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1052.920033] env[62814]: DEBUG nova.network.neutron [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1052.953215] env[62814]: DEBUG oslo_concurrency.lockutils [None req-16c9d8f3-1752-4730-a87a-4822595163f3 tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "508321ab-ce10-4953-a9e3-193b9975bec7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.800s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.973864] env[62814]: DEBUG oslo_vmware.api [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294323, 'name': PowerOnVM_Task, 'duration_secs': 0.572226} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.978820] env[62814]: DEBUG oslo_concurrency.lockutils [None req-910b1327-9909-40b5-aa34-046273771a9f tempest-InstanceActionsNegativeTestJSON-1315968954 tempest-InstanceActionsNegativeTestJSON-1315968954-project-member] Lock "a68901a9-9a9b-4127-bca9-64f98dfb151f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.903s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1052.984099] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1052.985616] env[62814]: DEBUG nova.compute.manager [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1052.985616] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639b505d-a1de-473e-89d6-eddae9a489d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.997811] env[62814]: DEBUG nova.policy [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '914cba82d20a43c28f31e33e09f61459', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '67cc9d091e094518a5ba7b78708698fe', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1053.030171] env[62814]: DEBUG oslo_vmware.api [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294324, 'name': PowerOnVM_Task, 'duration_secs': 0.563809} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.031172] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1053.031485] env[62814]: INFO nova.compute.manager [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Took 7.82 seconds to spawn the instance on the hypervisor. [ 1053.031766] env[62814]: DEBUG nova.compute.manager [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1053.033699] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68b3d4b-25b4-4aba-a0e0-1f78bd11f6ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.070788] env[62814]: DEBUG oslo_concurrency.lockutils [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "interface-885e0798-2fa6-4f6b-82be-517b6d1168d2-942dfa01-e165-41b2-8a4e-4d0d7dbb3f31" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1053.071207] env[62814]: DEBUG oslo_concurrency.lockutils [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-885e0798-2fa6-4f6b-82be-517b6d1168d2-942dfa01-e165-41b2-8a4e-4d0d7dbb3f31" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1053.203196] env[62814]: DEBUG oslo_vmware.api [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294327, 'name': ReconfigVM_Task, 'duration_secs': 0.196758} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.203514] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845740', 'volume_id': '4bf78821-2028-4edc-a038-062cb9d09c71', 'name': 'volume-4bf78821-2028-4edc-a038-062cb9d09c71', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '1e259ec6-d31a-453b-87e0-baa446665d56', 'attached_at': '', 'detached_at': '', 'volume_id': '4bf78821-2028-4edc-a038-062cb9d09c71', 'serial': '4bf78821-2028-4edc-a038-062cb9d09c71'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1053.229029] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294329, 'name': PowerOffVM_Task, 'duration_secs': 0.21844} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.229029] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.229029] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance '59d21ef7-df97-49ac-9329-4c18df6dd087' progress to 17 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1053.296831] env[62814]: DEBUG oslo_vmware.api [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Task: {'id': task-4294330, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180235} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.297162] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.297357] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1053.297979] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1053.297979] env[62814]: INFO nova.compute.manager [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1053.297979] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1053.298117] env[62814]: DEBUG nova.compute.manager [-] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1053.298211] env[62814]: DEBUG nova.network.neutron [-] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1053.340400] env[62814]: DEBUG nova.compute.manager [req-65ac20e8-3dff-415a-81f3-8cd7b062a22c req-e9ada3ec-eba7-4fa8-b3c5-e26ab455ae6d service nova] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Received event network-vif-deleted-3f437142-559d-418d-814f-10278dd5ed06 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1053.420884] env[62814]: DEBUG nova.compute.manager [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1053.509984] env[62814]: DEBUG oslo_concurrency.lockutils [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1053.564131] env[62814]: INFO nova.compute.manager [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Took 40.56 seconds to build instance. [ 1053.575302] env[62814]: DEBUG oslo_concurrency.lockutils [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.575614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1053.577075] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f9ed05-7ef4-4818-b1c5-f25249fa832d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.605544] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2489d64-4ac8-4cdc-a6b1-3ccee09d3aca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.674918] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Reconfiguring VM to detach interface {{(pid=62814) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1053.674918] env[62814]: DEBUG nova.network.neutron [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Successfully created port: b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1053.674918] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8bb4de8-a9e6-4455-b47a-e59b6ef1ddfe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.682269] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1053.682269] env[62814]: value = "task-4294331" [ 1053.682269] env[62814]: _type = "Task" [ 1053.682269] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.698132] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.739012] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1053.739729] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1053.739729] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1053.739729] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1053.739911] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1053.740061] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1053.740358] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1053.740550] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1053.740723] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1053.740889] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1053.741744] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1053.750306] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebb842fd-9059-4af8-aa59-be4504d377eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.772843] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1053.772843] env[62814]: value = "task-4294332" [ 1053.772843] env[62814]: _type = "Task" [ 1053.772843] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.778821] env[62814]: DEBUG nova.objects.instance [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lazy-loading 'flavor' on Instance uuid 1e259ec6-d31a-453b-87e0-baa446665d56 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1053.787552] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294332, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.998246] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0180053-faf9-4f16-a4a0-8b33c59718a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.007100] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30be33e2-b361-4bb3-9278-a019b75ef23f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.040510] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847313d5-0717-472c-9e5b-e705a40e5580 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.052987] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e25bb35-60db-43a6-84a3-ae67dd4ecd52 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.069331] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7cb4809e-ecc8-4539-9164-42e4febbda77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.081s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.069916] env[62814]: DEBUG nova.compute.provider_tree [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1054.084900] env[62814]: DEBUG nova.compute.manager [req-acf92de2-b337-4575-937e-ea241313aa30 req-33952798-2fe8-4a2c-9710-e1aab13b5394 service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Received event network-vif-deleted-22bf5740-3f2e-45ed-be98-ac62c70cfb3b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1054.084900] env[62814]: INFO nova.compute.manager [req-acf92de2-b337-4575-937e-ea241313aa30 req-33952798-2fe8-4a2c-9710-e1aab13b5394 service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Neutron deleted interface 22bf5740-3f2e-45ed-be98-ac62c70cfb3b; detaching it from the instance and deleting it from the info cache [ 1054.084900] env[62814]: DEBUG nova.network.neutron [req-acf92de2-b337-4575-937e-ea241313aa30 req-33952798-2fe8-4a2c-9710-e1aab13b5394 service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.175575] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.175834] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.177020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1054.177020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1054.177020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.178968] env[62814]: INFO nova.compute.manager [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Terminating instance [ 1054.192993] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.286512] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294332, 'name': ReconfigVM_Task, 'duration_secs': 0.320485} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.286902] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance '59d21ef7-df97-49ac-9329-4c18df6dd087' progress to 33 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1054.346801] env[62814]: DEBUG nova.network.neutron [-] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.442196] env[62814]: DEBUG nova.compute.manager [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1054.474343] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1054.474492] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.474669] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1054.475310] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.475310] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1054.475515] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1054.475648] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1054.475876] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1054.476169] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1054.476389] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1054.476765] env[62814]: DEBUG nova.virt.hardware [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1054.477637] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720a00cf-7ed2-4e4c-b83e-97abbe7bfeee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.488352] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c3bb9f-de48-46e6-ba27-c547580e1882 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.588238] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-454c33f1-2c2d-4fa6-8a45-eead70b65b7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.601724] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ce70491-9c6f-4013-ae20-88d30cbb45a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.618687] env[62814]: DEBUG nova.scheduler.client.report [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 132 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1054.618968] env[62814]: DEBUG nova.compute.provider_tree [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 132 to 133 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1054.619174] env[62814]: DEBUG nova.compute.provider_tree [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1054.641427] env[62814]: DEBUG nova.compute.manager [req-acf92de2-b337-4575-937e-ea241313aa30 req-33952798-2fe8-4a2c-9710-e1aab13b5394 service nova] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Detach interface failed, port_id=22bf5740-3f2e-45ed-be98-ac62c70cfb3b, reason: Instance c7cef7f3-11db-44e1-a454-98830b465b52 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1054.683806] env[62814]: DEBUG nova.compute.manager [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1054.684136] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1054.686074] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1c75793-0c9b-4170-9545-e763d2bad44c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.699231] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.700958] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1054.701234] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dff74d07-d8e1-412d-a15f-625b0a3af307 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.708683] env[62814]: DEBUG oslo_vmware.api [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 1054.708683] env[62814]: value = "task-4294333" [ 1054.708683] env[62814]: _type = "Task" [ 1054.708683] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.717746] env[62814]: DEBUG oslo_vmware.api [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294333, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.794682] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1054.794959] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.795199] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1054.795394] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.795550] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1054.795774] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1054.795989] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1054.796161] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1054.796323] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1054.796482] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1054.797232] env[62814]: DEBUG nova.virt.hardware [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1054.802373] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Reconfiguring VM instance instance-00000014 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1054.802859] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7096e5c4-7bb7-4526-a444-11fa512416af tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.301s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1054.803929] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1364dbd-416a-4123-84e6-fb497438e13b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.825183] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1054.825183] env[62814]: value = "task-4294334" [ 1054.825183] env[62814]: _type = "Task" [ 1054.825183] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.834914] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294334, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.849150] env[62814]: INFO nova.compute.manager [-] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Took 1.55 seconds to deallocate network for instance. [ 1055.127514] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.720s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.127514] env[62814]: DEBUG nova.compute.manager [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1055.135495] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.292s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.136178] env[62814]: INFO nova.compute.claims [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1055.198291] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.223090] env[62814]: DEBUG oslo_vmware.api [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294333, 'name': PowerOffVM_Task, 'duration_secs': 0.38777} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.224111] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1055.224301] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1055.224561] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f3aa4363-d914-471e-b7c4-3fe9418096d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.341570] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294334, 'name': ReconfigVM_Task, 'duration_secs': 0.270608} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.341570] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Reconfigured VM instance instance-00000014 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1055.342873] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13596149-48aa-46ae-ae2c-45b3deb668c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.365681] env[62814]: DEBUG oslo_concurrency.lockutils [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.378894] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Reconfiguring VM instance instance-00000014 to attach disk [datastore2] 59d21ef7-df97-49ac-9329-4c18df6dd087/59d21ef7-df97-49ac-9329-4c18df6dd087.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1055.380031] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ffc2a05a-72d2-49fd-be80-a0686772438c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.406019] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1055.406019] env[62814]: value = "task-4294336" [ 1055.406019] env[62814]: _type = "Task" [ 1055.406019] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.414226] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294336, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.433051] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "a07b964f-fe88-4c88-b6f4-61ed0973716c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.433411] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "a07b964f-fe88-4c88-b6f4-61ed0973716c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.433709] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "a07b964f-fe88-4c88-b6f4-61ed0973716c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.433937] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "a07b964f-fe88-4c88-b6f4-61ed0973716c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.434134] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "a07b964f-fe88-4c88-b6f4-61ed0973716c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.436468] env[62814]: INFO nova.compute.manager [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Terminating instance [ 1055.641062] env[62814]: DEBUG nova.compute.utils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1055.647774] env[62814]: DEBUG nova.compute.manager [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1055.647774] env[62814]: DEBUG nova.network.neutron [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1055.702612] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.758469] env[62814]: DEBUG nova.network.neutron [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Successfully updated port: b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1055.777267] env[62814]: DEBUG nova.policy [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1038a87615ae415ea4462635cb7d391d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '869db97f30f544f49e6a0c0e54bea463', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1055.910685] env[62814]: DEBUG nova.compute.manager [req-11a611f3-8cad-49b4-b5a6-31553bfbd3a7 req-d1e700c6-e191-4073-b9ec-c6a099c2503f service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Received event network-vif-plugged-b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1055.911214] env[62814]: DEBUG oslo_concurrency.lockutils [req-11a611f3-8cad-49b4-b5a6-31553bfbd3a7 req-d1e700c6-e191-4073-b9ec-c6a099c2503f service nova] Acquiring lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1055.911214] env[62814]: DEBUG oslo_concurrency.lockutils [req-11a611f3-8cad-49b4-b5a6-31553bfbd3a7 req-d1e700c6-e191-4073-b9ec-c6a099c2503f service nova] Lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1055.911310] env[62814]: DEBUG oslo_concurrency.lockutils [req-11a611f3-8cad-49b4-b5a6-31553bfbd3a7 req-d1e700c6-e191-4073-b9ec-c6a099c2503f service nova] Lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1055.911453] env[62814]: DEBUG nova.compute.manager [req-11a611f3-8cad-49b4-b5a6-31553bfbd3a7 req-d1e700c6-e191-4073-b9ec-c6a099c2503f service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] No waiting events found dispatching network-vif-plugged-b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1055.911667] env[62814]: WARNING nova.compute.manager [req-11a611f3-8cad-49b4-b5a6-31553bfbd3a7 req-d1e700c6-e191-4073-b9ec-c6a099c2503f service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Received unexpected event network-vif-plugged-b3b985f6-c18c-47da-a863-93d8e8a87f19 for instance with vm_state building and task_state spawning. [ 1055.921275] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294336, 'name': ReconfigVM_Task, 'duration_secs': 0.307886} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.921275] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Reconfigured VM instance instance-00000014 to attach disk [datastore2] 59d21ef7-df97-49ac-9329-4c18df6dd087/59d21ef7-df97-49ac-9329-4c18df6dd087.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1055.921275] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance '59d21ef7-df97-49ac-9329-4c18df6dd087' progress to 50 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1055.941784] env[62814]: DEBUG nova.compute.manager [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1055.942065] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1055.943052] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3968c76f-8c2f-4e2c-ae8b-9a9c7871da39 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.957028] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1055.958150] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51850e44-3447-42c6-884e-2e6415474868 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.971240] env[62814]: DEBUG oslo_vmware.api [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1055.971240] env[62814]: value = "task-4294337" [ 1055.971240] env[62814]: _type = "Task" [ 1055.971240] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.983347] env[62814]: DEBUG oslo_vmware.api [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294337, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.996596] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1055.996596] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1055.996596] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Deleting the datastore file [datastore2] f54b8bb9-69b7-4bb4-a82c-9f796050e719 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1055.996596] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d14384e3-0793-4ae8-893b-2403c6df19e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.006353] env[62814]: DEBUG oslo_vmware.api [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for the task: (returnval){ [ 1056.006353] env[62814]: value = "task-4294338" [ 1056.006353] env[62814]: _type = "Task" [ 1056.006353] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.017044] env[62814]: DEBUG oslo_vmware.api [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.152352] env[62814]: DEBUG nova.compute.manager [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1056.200106] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.263713] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.263713] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquired lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1056.263713] env[62814]: DEBUG nova.network.neutron [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1056.339077] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "7e849b23-226d-4a1d-a85c-cf96964ce034" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1056.339408] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1056.366084] env[62814]: DEBUG nova.network.neutron [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Successfully created port: 317f7973-158e-46c5-89bd-1e9555150bde {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1056.374416] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "1e259ec6-d31a-453b-87e0-baa446665d56" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1056.374990] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1056.375301] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "1e259ec6-d31a-453b-87e0-baa446665d56-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1056.375504] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1056.375676] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1056.379976] env[62814]: INFO nova.compute.manager [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Terminating instance [ 1056.434041] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e4fd8a8-6371-4b96-b382-393955a0e996 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.469352] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6bd11bb-295c-4563-a407-3edb607b4876 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.483852] env[62814]: DEBUG oslo_vmware.api [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294337, 'name': PowerOffVM_Task, 'duration_secs': 0.303084} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.500522] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1056.500741] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1056.501382] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance '59d21ef7-df97-49ac-9329-4c18df6dd087' progress to 67 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1056.508298] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-20eb8fd5-5a33-4a8f-86b1-3d6a31acfde9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.519921] env[62814]: DEBUG oslo_vmware.api [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Task: {'id': task-4294338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.35014} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.520737] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1056.521297] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1056.521719] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1056.522174] env[62814]: INFO nova.compute.manager [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Took 1.84 seconds to destroy the instance on the hypervisor. [ 1056.522447] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1056.524075] env[62814]: DEBUG nova.compute.manager [-] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1056.524075] env[62814]: DEBUG nova.network.neutron [-] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1056.575435] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1056.577502] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1056.577502] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleting the datastore file [datastore2] a07b964f-fe88-4c88-b6f4-61ed0973716c {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.577502] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2959f4b9-3ad5-4996-b71d-204d1799bb04 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.584532] env[62814]: DEBUG oslo_vmware.api [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1056.584532] env[62814]: value = "task-4294340" [ 1056.584532] env[62814]: _type = "Task" [ 1056.584532] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.599028] env[62814]: DEBUG oslo_vmware.api [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294340, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.698560] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.726840] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7c25ce-9222-4a79-beed-543cd3077952 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.735479] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25ce7d5c-20ef-484b-ae8f-69529cfb2415 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.773788] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3e5575-f9e9-40cc-8300-4df46b098fe0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.782744] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e7c264-af4f-4bfc-9b32-a9144a2d0016 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.798375] env[62814]: DEBUG nova.compute.provider_tree [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.838245] env[62814]: DEBUG nova.network.neutron [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1056.842360] env[62814]: DEBUG nova.compute.manager [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1056.886187] env[62814]: DEBUG nova.compute.manager [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1056.886187] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1056.886187] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52b582f0-e32f-4094-aa88-33a0125814af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.899717] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.900987] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2d97dc0a-564f-412d-abe9-19e375ec0175 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.912438] env[62814]: DEBUG oslo_vmware.api [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1056.912438] env[62814]: value = "task-4294341" [ 1056.912438] env[62814]: _type = "Task" [ 1056.912438] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.927016] env[62814]: DEBUG oslo_vmware.api [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294341, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.097385] env[62814]: DEBUG oslo_vmware.api [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294340, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.49845} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.097514] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.097852] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1057.098367] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1057.098597] env[62814]: INFO nova.compute.manager [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1057.098851] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1057.099414] env[62814]: DEBUG nova.compute.manager [-] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1057.099550] env[62814]: DEBUG nova.network.neutron [-] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1057.144337] env[62814]: DEBUG nova.network.neutron [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Port a00402c1-ebc9-40c0-93b2-26b6fbab4a68 binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1057.170251] env[62814]: DEBUG nova.compute.manager [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1057.203222] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.215200] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1057.215540] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1057.215718] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1057.219114] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1057.219114] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1057.219114] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1057.219114] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1057.219114] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1057.219339] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1057.219604] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1057.219892] env[62814]: DEBUG nova.virt.hardware [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1057.222557] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47087135-9442-422e-8021-808ea5d48d06 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.231326] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d0f3d52-3d3f-4968-9e30-71e1d2da6107 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.250821] env[62814]: DEBUG nova.network.neutron [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updating instance_info_cache with network_info: [{"id": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "address": "fa:16:3e:31:60:d0", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3b985f6-c1", "ovs_interfaceid": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.304144] env[62814]: DEBUG nova.scheduler.client.report [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1057.366993] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1057.428493] env[62814]: DEBUG oslo_vmware.api [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294341, 'name': PowerOffVM_Task, 'duration_secs': 0.401173} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.428493] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.428493] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1057.428493] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-150e6fc3-e42e-488a-b965-d1b2bf78468b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.499260] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1057.499260] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1057.499260] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleting the datastore file [datastore2] 1e259ec6-d31a-453b-87e0-baa446665d56 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.499260] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49f4423e-0306-4b5b-b4a2-03174905c045 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.507342] env[62814]: DEBUG oslo_vmware.api [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1057.507342] env[62814]: value = "task-4294343" [ 1057.507342] env[62814]: _type = "Task" [ 1057.507342] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.516345] env[62814]: DEBUG oslo_vmware.api [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294343, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.700524] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.754120] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Releasing lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1057.754485] env[62814]: DEBUG nova.compute.manager [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Instance network_info: |[{"id": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "address": "fa:16:3e:31:60:d0", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3b985f6-c1", "ovs_interfaceid": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1057.754921] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:60:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c883fb98-d172-4510-8cf4-07aafdf771af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3b985f6-c18c-47da-a863-93d8e8a87f19', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1057.763132] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Creating folder: Project (67cc9d091e094518a5ba7b78708698fe). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1057.763505] env[62814]: DEBUG nova.network.neutron [-] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.765277] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-807913ae-7bbf-4fe2-96c1-20b352a3413b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.781338] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Created folder: Project (67cc9d091e094518a5ba7b78708698fe) in parent group-v845547. [ 1057.781567] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Creating folder: Instances. Parent ref: group-v845804. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1057.781896] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1d9de33-e586-467f-a397-3b055184af2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.799272] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Created folder: Instances in parent group-v845804. [ 1057.799659] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1057.799889] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1057.800122] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac1a66d7-f51a-4573-9fe1-77eadd37aec4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.826806] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.693s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1057.827539] env[62814]: DEBUG nova.compute.manager [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1057.832505] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.726s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1057.834652] env[62814]: INFO nova.compute.claims [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1057.847358] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1057.847358] env[62814]: value = "task-4294346" [ 1057.847358] env[62814]: _type = "Task" [ 1057.847358] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.860122] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294346, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.971838] env[62814]: DEBUG nova.compute.manager [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Received event network-changed-b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1057.972179] env[62814]: DEBUG nova.compute.manager [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Refreshing instance network info cache due to event network-changed-b3b985f6-c18c-47da-a863-93d8e8a87f19. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1057.972508] env[62814]: DEBUG oslo_concurrency.lockutils [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] Acquiring lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.972754] env[62814]: DEBUG oslo_concurrency.lockutils [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] Acquired lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1057.973281] env[62814]: DEBUG nova.network.neutron [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Refreshing network info cache for port b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1058.018658] env[62814]: DEBUG oslo_vmware.api [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294343, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.371881} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.018958] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1058.019497] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1058.019497] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1058.019497] env[62814]: INFO nova.compute.manager [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1058.019770] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1058.020025] env[62814]: DEBUG nova.compute.manager [-] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1058.020143] env[62814]: DEBUG nova.network.neutron [-] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1058.141391] env[62814]: DEBUG oslo_concurrency.lockutils [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.141391] env[62814]: DEBUG oslo_concurrency.lockutils [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.141391] env[62814]: DEBUG nova.compute.manager [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1058.141391] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b86c9a-f528-498a-9690-17f16e506b3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.147820] env[62814]: DEBUG nova.compute.manager [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1058.148523] env[62814]: DEBUG nova.objects.instance [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lazy-loading 'flavor' on Instance uuid cfab341e-57a1-48b4-9b6b-ceecf28c223b {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1058.168082] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "59d21ef7-df97-49ac-9329-4c18df6dd087-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.168456] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.168561] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.199908] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.270306] env[62814]: INFO nova.compute.manager [-] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Took 1.75 seconds to deallocate network for instance. [ 1058.333736] env[62814]: DEBUG nova.compute.utils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1058.335814] env[62814]: DEBUG nova.compute.manager [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1058.335982] env[62814]: DEBUG nova.network.neutron [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1058.359075] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294346, 'name': CreateVM_Task, 'duration_secs': 0.337069} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.362130] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1058.363100] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.363324] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.364365] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1058.364888] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-628f478a-4b23-4aac-8c66-846e9d1142a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.373668] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1058.373668] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52297d2d-b52c-562e-524d-3b125602dd1f" [ 1058.373668] env[62814]: _type = "Task" [ 1058.373668] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.383357] env[62814]: DEBUG nova.network.neutron [-] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.384656] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52297d2d-b52c-562e-524d-3b125602dd1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.435926] env[62814]: DEBUG nova.compute.manager [req-c90f471e-6599-4727-b525-18b2775a60e7 req-a8d38bfe-3fc2-4e1e-8a3f-4f98695339b3 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Received event network-vif-plugged-317f7973-158e-46c5-89bd-1e9555150bde {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1058.436114] env[62814]: DEBUG oslo_concurrency.lockutils [req-c90f471e-6599-4727-b525-18b2775a60e7 req-a8d38bfe-3fc2-4e1e-8a3f-4f98695339b3 service nova] Acquiring lock "cd037f6e-fddd-4389-b6d0-144b798537bf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.436316] env[62814]: DEBUG oslo_concurrency.lockutils [req-c90f471e-6599-4727-b525-18b2775a60e7 req-a8d38bfe-3fc2-4e1e-8a3f-4f98695339b3 service nova] Lock "cd037f6e-fddd-4389-b6d0-144b798537bf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1058.436480] env[62814]: DEBUG oslo_concurrency.lockutils [req-c90f471e-6599-4727-b525-18b2775a60e7 req-a8d38bfe-3fc2-4e1e-8a3f-4f98695339b3 service nova] Lock "cd037f6e-fddd-4389-b6d0-144b798537bf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1058.436640] env[62814]: DEBUG nova.compute.manager [req-c90f471e-6599-4727-b525-18b2775a60e7 req-a8d38bfe-3fc2-4e1e-8a3f-4f98695339b3 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] No waiting events found dispatching network-vif-plugged-317f7973-158e-46c5-89bd-1e9555150bde {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1058.436800] env[62814]: WARNING nova.compute.manager [req-c90f471e-6599-4727-b525-18b2775a60e7 req-a8d38bfe-3fc2-4e1e-8a3f-4f98695339b3 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Received unexpected event network-vif-plugged-317f7973-158e-46c5-89bd-1e9555150bde for instance with vm_state building and task_state spawning. [ 1058.453627] env[62814]: DEBUG nova.policy [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7bec8437999a427e865f4c6723dec2c5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8b042739d3b4927a28b9729ec5dc08b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1058.463534] env[62814]: DEBUG nova.network.neutron [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Successfully updated port: 317f7973-158e-46c5-89bd-1e9555150bde {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1058.701246] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.776584] env[62814]: DEBUG nova.network.neutron [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updated VIF entry in instance network info cache for port b3b985f6-c18c-47da-a863-93d8e8a87f19. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1058.777171] env[62814]: DEBUG nova.network.neutron [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updating instance_info_cache with network_info: [{"id": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "address": "fa:16:3e:31:60:d0", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3b985f6-c1", "ovs_interfaceid": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.779611] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1058.840465] env[62814]: DEBUG nova.compute.manager [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1058.890017] env[62814]: INFO nova.compute.manager [-] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Took 1.79 seconds to deallocate network for instance. [ 1058.890017] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52297d2d-b52c-562e-524d-3b125602dd1f, 'name': SearchDatastore_Task, 'duration_secs': 0.03505} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.890017] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1058.890017] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1058.890459] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.890459] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.890459] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1058.893514] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f1213ac5-23fa-4acd-8efb-1ca85be1779f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.910946] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1058.910946] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1058.910946] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6ec908c-4f5c-4143-aae5-8edf74f7b5c2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.916868] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1058.916868] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522f2270-de3e-d343-a514-fece0e6dfdc5" [ 1058.916868] env[62814]: _type = "Task" [ 1058.916868] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.927110] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522f2270-de3e-d343-a514-fece0e6dfdc5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.969173] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "refresh_cache-cd037f6e-fddd-4389-b6d0-144b798537bf" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1058.969173] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "refresh_cache-cd037f6e-fddd-4389-b6d0-144b798537bf" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1058.969173] env[62814]: DEBUG nova.network.neutron [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1059.161902] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1059.161902] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f512a96-cb56-4335-a614-76dfdeb895cc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.171030] env[62814]: DEBUG oslo_vmware.api [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1059.171030] env[62814]: value = "task-4294347" [ 1059.171030] env[62814]: _type = "Task" [ 1059.171030] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.184224] env[62814]: DEBUG oslo_vmware.api [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294347, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.203490] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.247860] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.248060] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1059.248244] env[62814]: DEBUG nova.network.neutron [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1059.281314] env[62814]: DEBUG oslo_concurrency.lockutils [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] Releasing lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.281606] env[62814]: DEBUG nova.compute.manager [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Received event network-vif-deleted-20070e44-989b-4d3c-aab6-3de905aad547 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1059.281867] env[62814]: DEBUG nova.compute.manager [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Received event network-vif-deleted-08e04ba7-4665-4982-ac59-6412cc953d93 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1059.282050] env[62814]: INFO nova.compute.manager [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Neutron deleted interface 08e04ba7-4665-4982-ac59-6412cc953d93; detaching it from the instance and deleting it from the info cache [ 1059.282227] env[62814]: DEBUG nova.network.neutron [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.340220] env[62814]: DEBUG nova.network.neutron [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Successfully created port: 2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1059.351709] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6a95e0-069a-4534-bbec-4d20982fad5d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.358596] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c15114c-ffc0-4144-bf43-a26f93bfd8dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.392338] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eae8954-83e8-4da1-ab4c-2f4669464061 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.402291] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e71da2b-bad4-4c73-969f-eeeaec6bc8ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.407258] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1059.421455] env[62814]: DEBUG nova.compute.provider_tree [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.432374] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522f2270-de3e-d343-a514-fece0e6dfdc5, 'name': SearchDatastore_Task, 'duration_secs': 0.011856} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.433731] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-459f3fc4-010e-4c09-aed1-5ff891415eaf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.440537] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1059.440537] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e984b0-6195-4809-cd0f-020625731f38" [ 1059.440537] env[62814]: _type = "Task" [ 1059.440537] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.449830] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e984b0-6195-4809-cd0f-020625731f38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.558756] env[62814]: DEBUG nova.network.neutron [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1059.608059] env[62814]: DEBUG nova.network.neutron [-] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.685515] env[62814]: DEBUG oslo_vmware.api [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294347, 'name': PowerOffVM_Task, 'duration_secs': 0.432984} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.685799] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1059.687732] env[62814]: DEBUG nova.compute.manager [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1059.687732] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f57be147-fe13-4420-8d6a-a8cb4f6e99f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.704859] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.785065] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4cd7d345-2e3e-474f-a78f-b6cf51301e03 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.801020] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3720c6-99c2-4455-ac79-aa5a77d08db3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.836164] env[62814]: DEBUG nova.compute.manager [req-5a6f7920-a432-4185-83c6-0c0c90c51d58 req-f9f3cc4f-7adb-41f1-8437-e52f5179e8e9 service nova] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Detach interface failed, port_id=08e04ba7-4665-4982-ac59-6412cc953d93, reason: Instance a07b964f-fe88-4c88-b6f4-61ed0973716c could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1059.856079] env[62814]: DEBUG nova.compute.manager [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1059.894155] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1059.894321] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.894559] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1059.894805] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.896642] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1059.897356] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1059.897356] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1059.897356] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1059.897501] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1059.898229] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1059.898700] env[62814]: DEBUG nova.virt.hardware [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1059.899723] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9198871-30c8-41e1-b73c-5d1e4c8a4ca8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.908902] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bef7e1-e574-475f-a3e6-478bbd7a23b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.928013] env[62814]: DEBUG nova.scheduler.client.report [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1059.948901] env[62814]: DEBUG nova.network.neutron [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Updating instance_info_cache with network_info: [{"id": "317f7973-158e-46c5-89bd-1e9555150bde", "address": "fa:16:3e:63:41:7c", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap317f7973-15", "ovs_interfaceid": "317f7973-158e-46c5-89bd-1e9555150bde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.963913] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e984b0-6195-4809-cd0f-020625731f38, 'name': SearchDatastore_Task, 'duration_secs': 0.012332} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.967502] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1059.968301] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 0c6c598f-ef5a-4e91-b811-cd3d8f072647/0c6c598f-ef5a-4e91-b811-cd3d8f072647.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1059.968900] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db70fa7e-7b1c-42eb-a7fe-c5a17f8768f7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.979418] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1059.979418] env[62814]: value = "task-4294348" [ 1059.979418] env[62814]: _type = "Task" [ 1059.979418] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.991149] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294348, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.102704] env[62814]: DEBUG nova.network.neutron [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance_info_cache with network_info: [{"id": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "address": "fa:16:3e:93:37:3e", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa00402c1-eb", "ovs_interfaceid": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.107766] env[62814]: DEBUG nova.compute.manager [req-66a35152-393f-4894-a2df-60f235fababd req-48b3e01a-cb77-4f52-9b90-12908526222f service nova] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Received event network-vif-deleted-d9875d9c-cdba-4f00-b6a3-72216a3f9f22 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1060.113839] env[62814]: INFO nova.compute.manager [-] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Took 2.09 seconds to deallocate network for instance. [ 1060.206109] env[62814]: DEBUG oslo_vmware.api [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294331, 'name': ReconfigVM_Task, 'duration_secs': 6.361551} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.206429] env[62814]: DEBUG oslo_concurrency.lockutils [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1060.206671] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Reconfigured VM to detach interface {{(pid=62814) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1060.208979] env[62814]: DEBUG oslo_concurrency.lockutils [None req-158f258d-598e-4849-a473-05aeee969f80 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.070s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.440825] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.604s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1060.440825] env[62814]: DEBUG nova.compute.manager [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1060.444152] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.102s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1060.449704] env[62814]: INFO nova.compute.claims [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1060.467695] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "refresh_cache-cd037f6e-fddd-4389-b6d0-144b798537bf" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1060.468059] env[62814]: DEBUG nova.compute.manager [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Instance network_info: |[{"id": "317f7973-158e-46c5-89bd-1e9555150bde", "address": "fa:16:3e:63:41:7c", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap317f7973-15", "ovs_interfaceid": "317f7973-158e-46c5-89bd-1e9555150bde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1060.468850] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:63:41:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '317f7973-158e-46c5-89bd-1e9555150bde', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1060.481021] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1060.481021] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1060.481021] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f65b2cbb-4d6c-4e51-81ac-6a9c53b65656 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.508765] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294348, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.514982] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1060.514982] env[62814]: value = "task-4294349" [ 1060.514982] env[62814]: _type = "Task" [ 1060.514982] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.536160] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294349, 'name': CreateVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.564492] env[62814]: DEBUG nova.compute.manager [req-efa2f2b1-9fa6-4648-be56-a46af596aee2 req-81463c28-4f82-4612-943b-485465dff392 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Received event network-changed-317f7973-158e-46c5-89bd-1e9555150bde {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1060.564492] env[62814]: DEBUG nova.compute.manager [req-efa2f2b1-9fa6-4648-be56-a46af596aee2 req-81463c28-4f82-4612-943b-485465dff392 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Refreshing instance network info cache due to event network-changed-317f7973-158e-46c5-89bd-1e9555150bde. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1060.564737] env[62814]: DEBUG oslo_concurrency.lockutils [req-efa2f2b1-9fa6-4648-be56-a46af596aee2 req-81463c28-4f82-4612-943b-485465dff392 service nova] Acquiring lock "refresh_cache-cd037f6e-fddd-4389-b6d0-144b798537bf" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.564902] env[62814]: DEBUG oslo_concurrency.lockutils [req-efa2f2b1-9fa6-4648-be56-a46af596aee2 req-81463c28-4f82-4612-943b-485465dff392 service nova] Acquired lock "refresh_cache-cd037f6e-fddd-4389-b6d0-144b798537bf" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1060.564984] env[62814]: DEBUG nova.network.neutron [req-efa2f2b1-9fa6-4648-be56-a46af596aee2 req-81463c28-4f82-4612-943b-485465dff392 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Refreshing network info cache for port 317f7973-158e-46c5-89bd-1e9555150bde {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1060.610375] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1060.623779] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1060.920267] env[62814]: DEBUG nova.objects.instance [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lazy-loading 'flavor' on Instance uuid cfab341e-57a1-48b4-9b6b-ceecf28c223b {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1060.955405] env[62814]: DEBUG nova.compute.utils [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1060.960023] env[62814]: DEBUG nova.compute.manager [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1060.960023] env[62814]: DEBUG nova.network.neutron [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1061.009619] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294348, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.823585} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.010314] env[62814]: DEBUG nova.policy [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b976d86c39f423f8fe03a06ea3ecfc2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0060ccfb18f34fbca1ab765134395bb8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1061.012920] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 0c6c598f-ef5a-4e91-b811-cd3d8f072647/0c6c598f-ef5a-4e91-b811-cd3d8f072647.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1061.012920] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1061.012920] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-062933b9-31c3-4f03-8633-53fe0506ccf1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.028117] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294349, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.031211] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1061.031211] env[62814]: value = "task-4294350" [ 1061.031211] env[62814]: _type = "Task" [ 1061.031211] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.153021] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e0e99e-4174-400b-8a80-57df46be522e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.184173] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82c2120-2a30-4cb9-a7ea-dd6c044e1be7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.195301] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance '59d21ef7-df97-49ac-9329-4c18df6dd087' progress to 83 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1061.427638] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "refresh_cache-cfab341e-57a1-48b4-9b6b-ceecf28c223b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.427938] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquired lock "refresh_cache-cfab341e-57a1-48b4-9b6b-ceecf28c223b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.428453] env[62814]: DEBUG nova.network.neutron [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1061.428672] env[62814]: DEBUG nova.objects.instance [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lazy-loading 'info_cache' on Instance uuid cfab341e-57a1-48b4-9b6b-ceecf28c223b {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.455394] env[62814]: DEBUG nova.network.neutron [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Successfully created port: dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1061.465834] env[62814]: DEBUG nova.compute.manager [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1061.500025] env[62814]: DEBUG nova.network.neutron [req-efa2f2b1-9fa6-4648-be56-a46af596aee2 req-81463c28-4f82-4612-943b-485465dff392 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Updated VIF entry in instance network info cache for port 317f7973-158e-46c5-89bd-1e9555150bde. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1061.500025] env[62814]: DEBUG nova.network.neutron [req-efa2f2b1-9fa6-4648-be56-a46af596aee2 req-81463c28-4f82-4612-943b-485465dff392 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Updating instance_info_cache with network_info: [{"id": "317f7973-158e-46c5-89bd-1e9555150bde", "address": "fa:16:3e:63:41:7c", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap317f7973-15", "ovs_interfaceid": "317f7973-158e-46c5-89bd-1e9555150bde", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.529280] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294349, 'name': CreateVM_Task, 'duration_secs': 0.625934} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.530306] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1061.530306] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.530475] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.530849] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1061.531221] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94464887-4817-456e-9107-b7f9618ee423 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.550649] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1061.550649] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5201a07d-1f86-d7c3-5d28-7f2da1a00358" [ 1061.550649] env[62814]: _type = "Task" [ 1061.550649] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.551258] env[62814]: DEBUG oslo_concurrency.lockutils [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.551468] env[62814]: DEBUG oslo_concurrency.lockutils [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.551667] env[62814]: DEBUG nova.network.neutron [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1061.564657] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294350, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.359766} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.567531] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1061.574698] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5d4cb9-b6d2-4a7e-b055-acfe3181e931 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.584490] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5201a07d-1f86-d7c3-5d28-7f2da1a00358, 'name': SearchDatastore_Task, 'duration_secs': 0.016377} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.585589] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1061.585943] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.586170] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.586336] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1061.586540] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.586883] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a2c6e81-e458-48b0-9a13-b62323329c4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.611327] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 0c6c598f-ef5a-4e91-b811-cd3d8f072647/0c6c598f-ef5a-4e91-b811-cd3d8f072647.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1061.615967] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58e30ff2-4c0c-4353-90a5-53648fcd86e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.636185] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1061.636416] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1061.637957] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-216896fd-18a2-4712-8837-6540ca7ff909 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.645531] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1061.645531] env[62814]: value = "task-4294351" [ 1061.645531] env[62814]: _type = "Task" [ 1061.645531] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.647140] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1061.647140] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528b0e8b-556f-4695-021c-cc670361f4da" [ 1061.647140] env[62814]: _type = "Task" [ 1061.647140] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.662766] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294351, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.665800] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528b0e8b-556f-4695-021c-cc670361f4da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.701947] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1061.702107] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e59c0148-0a88-4134-a02c-6a506c3f6884 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.714682] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1061.714682] env[62814]: value = "task-4294352" [ 1061.714682] env[62814]: _type = "Task" [ 1061.714682] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.728320] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294352, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.839820] env[62814]: DEBUG nova.network.neutron [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Successfully updated port: 2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1061.938382] env[62814]: DEBUG nova.objects.base [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1061.973386] env[62814]: INFO nova.virt.block_device [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Booting with volume 6f2eddd0-4334-4e22-b195-b7deaa539235 at /dev/sda [ 1062.004833] env[62814]: DEBUG oslo_concurrency.lockutils [req-efa2f2b1-9fa6-4648-be56-a46af596aee2 req-81463c28-4f82-4612-943b-485465dff392 service nova] Releasing lock "refresh_cache-cd037f6e-fddd-4389-b6d0-144b798537bf" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1062.025345] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea235ef-e4e7-4ba7-9fce-0516fdcf9616 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.043270] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a43d0a7-5945-4375-9b02-4d2781406ed7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.047536] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c85e1cff-659d-49c1-87b0-34ad93f89df7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.087925] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64aa7b49-55f8-480f-866d-691b3fd87796 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.094648] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b48f33f2-a625-4e89-b89a-fb88bca4bb69 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.114218] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371839b5-651b-4cae-ade1-9dc1f247f9d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.141196] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21eac509-36ce-446f-ae17-71202778c7c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.143619] env[62814]: DEBUG nova.compute.provider_tree [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.159442] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f257a26a-c6c8-4f5f-a7f8-bf936c45aadc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.181220] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528b0e8b-556f-4695-021c-cc670361f4da, 'name': SearchDatastore_Task, 'duration_secs': 0.033272} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.181681] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294351, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.183426] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b50bf3b9-ccc6-48e2-8f6f-dfcd0fefe027 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.195809] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1062.195809] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aae043-231a-3db6-f7b7-6e308c442dea" [ 1062.195809] env[62814]: _type = "Task" [ 1062.195809] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.213289] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e8e79f-74b5-4886-9fef-c9496aedb4ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.220062] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aae043-231a-3db6-f7b7-6e308c442dea, 'name': SearchDatastore_Task, 'duration_secs': 0.020597} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.223775] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1062.224093] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] cd037f6e-fddd-4389-b6d0-144b798537bf/cd037f6e-fddd-4389-b6d0-144b798537bf.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1062.226662] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2340a283-948d-4a34-9792-121b75cd828a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.229995] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5ceade-01d4-41f2-9496-cdfeb688e73d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.236567] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294352, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.242662] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1062.242662] env[62814]: value = "task-4294353" [ 1062.242662] env[62814]: _type = "Task" [ 1062.242662] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.253376] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294353, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.256323] env[62814]: DEBUG nova.virt.block_device [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Updating existing volume attachment record: e73b094c-8fac-409a-b8ab-24f1db58d784 {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1062.260447] env[62814]: DEBUG oslo_concurrency.lockutils [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.260447] env[62814]: DEBUG oslo_concurrency.lockutils [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.260447] env[62814]: DEBUG oslo_concurrency.lockutils [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "885e0798-2fa6-4f6b-82be-517b6d1168d2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.260447] env[62814]: DEBUG oslo_concurrency.lockutils [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "885e0798-2fa6-4f6b-82be-517b6d1168d2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.260447] env[62814]: DEBUG oslo_concurrency.lockutils [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "885e0798-2fa6-4f6b-82be-517b6d1168d2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1062.267959] env[62814]: INFO nova.compute.manager [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Terminating instance [ 1062.346110] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.346343] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1062.346646] env[62814]: DEBUG nova.network.neutron [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1062.586522] env[62814]: INFO nova.network.neutron [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Port 942dfa01-e165-41b2-8a4e-4d0d7dbb3f31 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1062.587055] env[62814]: DEBUG nova.network.neutron [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updating instance_info_cache with network_info: [{"id": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "address": "fa:16:3e:ee:a8:e0", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap75b18381-f5", "ovs_interfaceid": "75b18381-f5ef-4ce2-ab71-634c787eaffe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.648307] env[62814]: DEBUG nova.compute.manager [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Received event network-vif-plugged-2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1062.648623] env[62814]: DEBUG oslo_concurrency.lockutils [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] Acquiring lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.648931] env[62814]: DEBUG oslo_concurrency.lockutils [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] Lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.649411] env[62814]: DEBUG oslo_concurrency.lockutils [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] Lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1062.649497] env[62814]: DEBUG nova.compute.manager [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] No waiting events found dispatching network-vif-plugged-2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1062.650068] env[62814]: WARNING nova.compute.manager [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Received unexpected event network-vif-plugged-2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f for instance with vm_state building and task_state spawning. [ 1062.650347] env[62814]: DEBUG nova.compute.manager [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Received event network-changed-2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1062.650450] env[62814]: DEBUG nova.compute.manager [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Refreshing instance network info cache due to event network-changed-2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1062.651155] env[62814]: DEBUG oslo_concurrency.lockutils [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] Acquiring lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.652551] env[62814]: DEBUG nova.scheduler.client.report [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1062.671128] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294351, 'name': ReconfigVM_Task, 'duration_secs': 0.741117} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.671998] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 0c6c598f-ef5a-4e91-b811-cd3d8f072647/0c6c598f-ef5a-4e91-b811-cd3d8f072647.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1062.673169] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c376ad72-f526-4210-ad9a-b96e447b48d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.683272] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1062.683272] env[62814]: value = "task-4294354" [ 1062.683272] env[62814]: _type = "Task" [ 1062.683272] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.696872] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294354, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.732366] env[62814]: DEBUG oslo_vmware.api [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294352, 'name': PowerOnVM_Task, 'duration_secs': 0.609848} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.732667] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1062.732955] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4572969d-1f49-4f1f-b683-3dface1b3e51 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance '59d21ef7-df97-49ac-9329-4c18df6dd087' progress to 100 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1062.754909] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294353, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.768615] env[62814]: DEBUG nova.compute.manager [req-cd835e80-c506-46cc-b77f-8f4cc94fc984 req-7204bfc8-345f-441f-a57a-d29bff3c7e5e service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Received event network-vif-deleted-942dfa01-e165-41b2-8a4e-4d0d7dbb3f31 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1062.779772] env[62814]: DEBUG nova.compute.manager [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1062.779772] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1062.779772] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f35d47-d58b-4fbe-a77d-785e185abc29 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.788201] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1062.788201] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-204c7ff6-3498-4bdd-a08b-b40007679424 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.798290] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquiring lock "4429ff0c-5353-4346-8bef-2c45d251f1ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1062.798807] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lock "4429ff0c-5353-4346-8bef-2c45d251f1ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1062.800209] env[62814]: DEBUG oslo_vmware.api [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1062.800209] env[62814]: value = "task-4294355" [ 1062.800209] env[62814]: _type = "Task" [ 1062.800209] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.814213] env[62814]: DEBUG nova.network.neutron [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Updating instance_info_cache with network_info: [{"id": "496dd017-79fb-4bcc-8fdb-32ab21f6b676", "address": "fa:16:3e:22:4b:29", "network": {"id": "c3a9b4ad-b825-451e-ad5a-4e0331bb2d16", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-1261267397-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2e8761ca6d3444b2a94b18ef8628f4f4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap496dd017-79", "ovs_interfaceid": "496dd017-79fb-4bcc-8fdb-32ab21f6b676", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.818450] env[62814]: DEBUG oslo_vmware.api [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.912911] env[62814]: DEBUG nova.network.neutron [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1063.092786] env[62814]: DEBUG oslo_concurrency.lockutils [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-885e0798-2fa6-4f6b-82be-517b6d1168d2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.101610] env[62814]: DEBUG nova.network.neutron [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Updating instance_info_cache with network_info: [{"id": "2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f", "address": "fa:16:3e:b5:d9:f0", "network": {"id": "dbc702b4-6191-4068-a1d6-4c03cc349962", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1378843546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d8b042739d3b4927a28b9729ec5dc08b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2abec6f6-ac", "ovs_interfaceid": "2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.160157] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.715s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.160157] env[62814]: DEBUG nova.compute.manager [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1063.164055] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.011s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1063.164055] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.165776] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 18.891s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1063.197837] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294354, 'name': Rename_Task, 'duration_secs': 0.262898} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.198877] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1063.198877] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-59574a25-6886-426f-9228-ab9ebe5ffc6c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.203655] env[62814]: INFO nova.scheduler.client.report [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted allocations for instance 4ebc3280-64a7-4e9b-8385-2128357d3422 [ 1063.211640] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1063.211640] env[62814]: value = "task-4294356" [ 1063.211640] env[62814]: _type = "Task" [ 1063.211640] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.222391] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294356, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.256122] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294353, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.643765} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.256932] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] cd037f6e-fddd-4389-b6d0-144b798537bf/cd037f6e-fddd-4389-b6d0-144b798537bf.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1063.258884] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.259957] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75f93be6-0c06-4fb1-a056-f0e100dc4439 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.267352] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1063.267352] env[62814]: value = "task-4294357" [ 1063.267352] env[62814]: _type = "Task" [ 1063.267352] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.278716] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294357, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.301877] env[62814]: DEBUG nova.compute.manager [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1063.315208] env[62814]: DEBUG oslo_vmware.api [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294355, 'name': PowerOffVM_Task, 'duration_secs': 0.212185} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.315208] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1063.315208] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1063.315330] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b091244f-0ace-4735-aead-9c9aae277333 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.320555] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Releasing lock "refresh_cache-cfab341e-57a1-48b4-9b6b-ceecf28c223b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.341074] env[62814]: DEBUG nova.network.neutron [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Successfully updated port: dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1063.382060] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1063.382188] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1063.382335] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleting the datastore file [datastore2] 885e0798-2fa6-4f6b-82be-517b6d1168d2 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1063.382892] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94b8f6f0-1974-49b7-8e07-a59a06288b62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.392231] env[62814]: DEBUG oslo_vmware.api [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1063.392231] env[62814]: value = "task-4294359" [ 1063.392231] env[62814]: _type = "Task" [ 1063.392231] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.401754] env[62814]: DEBUG oslo_vmware.api [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294359, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.594695] env[62814]: DEBUG oslo_concurrency.lockutils [None req-994d5fb3-033b-48a8-9596-afd3217abb82 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-885e0798-2fa6-4f6b-82be-517b6d1168d2-942dfa01-e165-41b2-8a4e-4d0d7dbb3f31" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.523s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.603610] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1063.604070] env[62814]: DEBUG nova.compute.manager [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Instance network_info: |[{"id": "2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f", "address": "fa:16:3e:b5:d9:f0", "network": {"id": "dbc702b4-6191-4068-a1d6-4c03cc349962", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1378843546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d8b042739d3b4927a28b9729ec5dc08b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2abec6f6-ac", "ovs_interfaceid": "2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1063.604730] env[62814]: DEBUG oslo_concurrency.lockutils [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] Acquired lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1063.605438] env[62814]: DEBUG nova.network.neutron [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Refreshing network info cache for port 2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1063.606681] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:d9:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e6f11c0d-c73a-47f5-b02e-47bff48da0e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1063.616641] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1063.617657] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1063.620507] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73ac4159-feed-46dd-be68-0460afed23dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.642399] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1063.642399] env[62814]: value = "task-4294360" [ 1063.642399] env[62814]: _type = "Task" [ 1063.642399] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.651335] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294360, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.669413] env[62814]: DEBUG nova.compute.utils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1063.673474] env[62814]: DEBUG nova.compute.manager [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1063.673933] env[62814]: DEBUG nova.network.neutron [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1063.713258] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fce2b346-1e93-4009-9eb2-87f225d76da5 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "4ebc3280-64a7-4e9b-8385-2128357d3422" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.197s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1063.729426] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294356, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.738544] env[62814]: DEBUG nova.policy [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '93a43d89fcb34e4eaec597b28caf45fc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1c7dab0a5e874c3aaa57685a130460fc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1063.779462] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294357, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.830745] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1063.844532] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Acquiring lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.844532] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Acquired lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1063.844682] env[62814]: DEBUG nova.network.neutron [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1063.905637] env[62814]: DEBUG oslo_vmware.api [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294359, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161105} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.906100] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1063.906427] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1063.909479] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1063.909479] env[62814]: INFO nova.compute.manager [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1063.909479] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1063.909479] env[62814]: DEBUG nova.compute.manager [-] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1063.909479] env[62814]: DEBUG nova.network.neutron [-] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1064.108373] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5da0cd-9bca-4a79-8db2-b1b813b430cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.116841] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0474df1d-c007-4b9c-9ba5-8a121f5a87a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.163534] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0376d2fa-5345-440b-acb6-03e7a57385f3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.167641] env[62814]: DEBUG nova.network.neutron [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Successfully created port: d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1064.180302] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e836f364-76ca-4f02-88e6-72442d9fa0cc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.183486] env[62814]: DEBUG nova.compute.manager [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1064.186252] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294360, 'name': CreateVM_Task, 'duration_secs': 0.407066} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.186470] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1064.188033] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.188273] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.188630] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1064.188915] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a99d9f38-1b35-43ff-9ec4-6288ccf5488c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.200779] env[62814]: DEBUG nova.compute.provider_tree [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.213180] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1064.213180] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5236e4cc-ec88-31c6-fd88-583601eb86a2" [ 1064.213180] env[62814]: _type = "Task" [ 1064.213180] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.227396] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5236e4cc-ec88-31c6-fd88-583601eb86a2, 'name': SearchDatastore_Task, 'duration_secs': 0.010861} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.231151] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1064.231485] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1064.231785] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.231945] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1064.232139] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1064.232421] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294356, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.232630] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d2f5d9f-4d15-443b-bf27-062670028deb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.243462] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1064.243649] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1064.244678] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8387ed9e-16a5-4b01-ae79-a77949bf0f77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.258900] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1064.258900] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525eb26a-9c6d-f0c9-c3a4-0311c50819ef" [ 1064.258900] env[62814]: _type = "Task" [ 1064.258900] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.281013] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525eb26a-9c6d-f0c9-c3a4-0311c50819ef, 'name': SearchDatastore_Task, 'duration_secs': 0.011713} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.283438] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae98a636-5a69-4f20-bb1e-382897c16620 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.289394] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294357, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.753483} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.290098] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1064.290926] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3677f3-8f3e-4eb9-92eb-170ab281fda8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.295085] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1064.295085] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ed43ed-91e9-149d-9733-d4024709d524" [ 1064.295085] env[62814]: _type = "Task" [ 1064.295085] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.318925] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] cd037f6e-fddd-4389-b6d0-144b798537bf/cd037f6e-fddd-4389-b6d0-144b798537bf.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.320026] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fbac93b-b490-43c1-bf91-78a6d01afa04 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.341155] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1064.341631] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ed43ed-91e9-149d-9733-d4024709d524, 'name': SearchDatastore_Task, 'duration_secs': 0.010988} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.341952] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe832acf-8ae1-4ae3-9b2c-26a8a3f4a1a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.344134] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1064.344548] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3/19b53797-6fc9-4bb0-ab3f-9d2941b11cb3.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1064.344658] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d13b0b58-17f9-414c-a85f-a5ac2ea2d8fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.352467] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1064.352467] env[62814]: value = "task-4294361" [ 1064.352467] env[62814]: _type = "Task" [ 1064.352467] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.357069] env[62814]: DEBUG oslo_vmware.api [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1064.357069] env[62814]: value = "task-4294362" [ 1064.357069] env[62814]: _type = "Task" [ 1064.357069] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.366069] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1064.366069] env[62814]: value = "task-4294363" [ 1064.366069] env[62814]: _type = "Task" [ 1064.366069] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.366311] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294361, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.372763] env[62814]: DEBUG oslo_vmware.api [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294362, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.378177] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294363, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.386755] env[62814]: DEBUG nova.compute.manager [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1064.387423] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1064.387709] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.387936] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1064.388207] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.388423] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1064.388630] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1064.388939] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1064.389149] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1064.389449] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1064.389640] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1064.390051] env[62814]: DEBUG nova.virt.hardware [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1064.391030] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086a5b98-ec6e-4b85-85d6-e479067d5cd5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.400717] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a74aa08-a2cb-4cd6-a51a-d991302d74a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.433564] env[62814]: DEBUG nova.network.neutron [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1064.708199] env[62814]: DEBUG nova.scheduler.client.report [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1064.728309] env[62814]: DEBUG oslo_vmware.api [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294356, 'name': PowerOnVM_Task, 'duration_secs': 1.37624} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.729501] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1064.729778] env[62814]: INFO nova.compute.manager [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Took 10.29 seconds to spawn the instance on the hypervisor. [ 1064.730042] env[62814]: DEBUG nova.compute.manager [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1064.731328] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7bc0aa5-f0cc-4864-a3e7-0e6143a652e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.735446] env[62814]: DEBUG nova.network.neutron [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Updated VIF entry in instance network info cache for port 2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1064.735813] env[62814]: DEBUG nova.network.neutron [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Updating instance_info_cache with network_info: [{"id": "2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f", "address": "fa:16:3e:b5:d9:f0", "network": {"id": "dbc702b4-6191-4068-a1d6-4c03cc349962", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1378843546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d8b042739d3b4927a28b9729ec5dc08b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2abec6f6-ac", "ovs_interfaceid": "2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.850204] env[62814]: DEBUG nova.network.neutron [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Updating instance_info_cache with network_info: [{"id": "dec4b0c4-68db-48fc-84e2-85becf8933bc", "address": "fa:16:3e:f5:e8:b4", "network": {"id": "afa6767f-5aa7-45a5-b700-6237ed1b36c5", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-40244317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0060ccfb18f34fbca1ab765134395bb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdec4b0c4-68", "ovs_interfaceid": "dec4b0c4-68db-48fc-84e2-85becf8933bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.867070] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294361, 'name': ReconfigVM_Task, 'duration_secs': 0.42671} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.867876] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Reconfigured VM instance instance-00000059 to attach disk [datastore2] cd037f6e-fddd-4389-b6d0-144b798537bf/cd037f6e-fddd-4389-b6d0-144b798537bf.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.871946] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a20c3a6-3973-4bec-84bf-1584479afe98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.878083] env[62814]: DEBUG oslo_vmware.api [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294362, 'name': PowerOnVM_Task} progress is 96%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.885381] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294363, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.887508] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1064.887508] env[62814]: value = "task-4294364" [ 1064.887508] env[62814]: _type = "Task" [ 1064.887508] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.898058] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294364, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.109087] env[62814]: DEBUG nova.compute.manager [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Received event network-vif-plugged-dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1065.110909] env[62814]: DEBUG oslo_concurrency.lockutils [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] Acquiring lock "d5df1d0e-55e6-452e-882a-053a83250b39-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.110909] env[62814]: DEBUG oslo_concurrency.lockutils [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] Lock "d5df1d0e-55e6-452e-882a-053a83250b39-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.110909] env[62814]: DEBUG oslo_concurrency.lockutils [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] Lock "d5df1d0e-55e6-452e-882a-053a83250b39-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.110909] env[62814]: DEBUG nova.compute.manager [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] No waiting events found dispatching network-vif-plugged-dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1065.110909] env[62814]: WARNING nova.compute.manager [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Received unexpected event network-vif-plugged-dec4b0c4-68db-48fc-84e2-85becf8933bc for instance with vm_state building and task_state spawning. [ 1065.110909] env[62814]: DEBUG nova.compute.manager [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Received event network-changed-dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1065.110909] env[62814]: DEBUG nova.compute.manager [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Refreshing instance network info cache due to event network-changed-dec4b0c4-68db-48fc-84e2-85becf8933bc. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1065.110909] env[62814]: DEBUG oslo_concurrency.lockutils [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] Acquiring lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.156172] env[62814]: DEBUG nova.compute.manager [req-7c6ea828-8d04-4d44-85bc-6ecb0b2f29fe req-b8f423a2-a37e-41c2-89d4-31dd6952fe67 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Received event network-vif-deleted-75b18381-f5ef-4ce2-ab71-634c787eaffe {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1065.156172] env[62814]: INFO nova.compute.manager [req-7c6ea828-8d04-4d44-85bc-6ecb0b2f29fe req-b8f423a2-a37e-41c2-89d4-31dd6952fe67 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Neutron deleted interface 75b18381-f5ef-4ce2-ab71-634c787eaffe; detaching it from the instance and deleting it from the info cache [ 1065.156172] env[62814]: DEBUG nova.network.neutron [req-7c6ea828-8d04-4d44-85bc-6ecb0b2f29fe req-b8f423a2-a37e-41c2-89d4-31dd6952fe67 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.196095] env[62814]: DEBUG nova.compute.manager [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1065.233087] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1065.233868] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1065.233868] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1065.233868] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1065.234141] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1065.234387] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1065.234663] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1065.234923] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1065.235102] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1065.235327] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1065.235570] env[62814]: DEBUG nova.virt.hardware [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1065.236844] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b911eb-193f-47eb-9a1f-cc558934adbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.241022] env[62814]: DEBUG oslo_concurrency.lockutils [req-13aa87bf-ec13-467b-8f72-fd399d26e23a req-0d707497-a039-4494-a2db-c8f2a3d85c4c service nova] Releasing lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.247656] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16468e2c-8878-4385-96b0-a6194573a987 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.270025] env[62814]: INFO nova.compute.manager [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Took 45.75 seconds to build instance. [ 1065.353604] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Releasing lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1065.353986] env[62814]: DEBUG nova.compute.manager [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Instance network_info: |[{"id": "dec4b0c4-68db-48fc-84e2-85becf8933bc", "address": "fa:16:3e:f5:e8:b4", "network": {"id": "afa6767f-5aa7-45a5-b700-6237ed1b36c5", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-40244317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0060ccfb18f34fbca1ab765134395bb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdec4b0c4-68", "ovs_interfaceid": "dec4b0c4-68db-48fc-84e2-85becf8933bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1065.354342] env[62814]: DEBUG oslo_concurrency.lockutils [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] Acquired lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1065.354528] env[62814]: DEBUG nova.network.neutron [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Refreshing network info cache for port dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1065.355986] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:e8:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e350f83a-f581-4e10-ac16-0b0f7bfd3d38', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dec4b0c4-68db-48fc-84e2-85becf8933bc', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1065.363837] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Creating folder: Project (0060ccfb18f34fbca1ab765134395bb8). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1065.365050] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d49b93e5-0161-4fbd-bdb0-ca34fb3b7faa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.380820] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294363, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550469} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.384031] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3/19b53797-6fc9-4bb0-ab3f-9d2941b11cb3.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1065.384031] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1065.384223] env[62814]: DEBUG oslo_vmware.api [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294362, 'name': PowerOnVM_Task, 'duration_secs': 0.608891} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.385476] env[62814]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1065.385591] env[62814]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62814) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1065.385801] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a8f5397a-1a24-44bf-891f-fdbe38742d7d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.387711] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.387917] env[62814]: DEBUG nova.compute.manager [None req-2db60641-8b32-4551-8dbb-45ebb05daf3d tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1065.388235] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Folder already exists: Project (0060ccfb18f34fbca1ab765134395bb8). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1065.388416] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Creating folder: Instances. Parent ref: group-v845775. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1065.389191] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ac0793-cd21-42fc-ad3f-dd26f3ee3bd0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.392226] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb7ff46a-e855-4641-9a26-9c08877296a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.399736] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1065.399736] env[62814]: value = "task-4294366" [ 1065.399736] env[62814]: _type = "Task" [ 1065.399736] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.411236] env[62814]: DEBUG nova.network.neutron [-] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.412976] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294364, 'name': Rename_Task, 'duration_secs': 0.464982} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.415190] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1065.415508] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Created folder: Instances in parent group-v845775. [ 1065.415745] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1065.416365] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d50e5bd9-b9a5-46f6-a5bf-6391e329bc7c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.418050] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1065.421562] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68b5272e-2ee8-41d7-a816-0ba9ac710eec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.439194] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294366, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.450022] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1065.450022] env[62814]: value = "task-4294369" [ 1065.450022] env[62814]: _type = "Task" [ 1065.450022] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.450022] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1065.450022] env[62814]: value = "task-4294368" [ 1065.450022] env[62814]: _type = "Task" [ 1065.450022] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.467092] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294369, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.470927] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294368, 'name': CreateVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.627786] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.628049] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.658521] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6cc39b33-3274-492c-ab75-b7f4b55bffe7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.670331] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-511f02ee-6d78-4aa8-be7e-9b196c9f2736 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.709762] env[62814]: DEBUG nova.compute.manager [req-7c6ea828-8d04-4d44-85bc-6ecb0b2f29fe req-b8f423a2-a37e-41c2-89d4-31dd6952fe67 service nova] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Detach interface failed, port_id=75b18381-f5ef-4ce2-ab71-634c787eaffe, reason: Instance 885e0798-2fa6-4f6b-82be-517b6d1168d2 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1065.728075] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.562s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.728237] env[62814]: DEBUG nova.compute.manager [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62814) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5376}} [ 1065.731136] env[62814]: DEBUG oslo_concurrency.lockutils [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.803s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.731356] env[62814]: DEBUG nova.objects.instance [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lazy-loading 'resources' on Instance uuid 88ec5aba-f38d-4c75-af29-e3df3aa49640 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.776368] env[62814]: DEBUG oslo_concurrency.lockutils [None req-069477b6-c242-446c-9027-0f5a74dabf9d tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.265s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1065.777185] env[62814]: DEBUG oslo_concurrency.lockutils [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "59d21ef7-df97-49ac-9329-4c18df6dd087" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1065.777286] env[62814]: DEBUG oslo_concurrency.lockutils [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1065.777414] env[62814]: DEBUG nova.compute.manager [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Going to confirm migration 4 {{(pid=62814) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 1065.912050] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294366, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087798} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.912340] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1065.913188] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91629e3e-28ad-4a42-b511-24cd6ce8c1b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.915858] env[62814]: INFO nova.compute.manager [-] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Took 2.01 seconds to deallocate network for instance. [ 1065.953642] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3/19b53797-6fc9-4bb0-ab3f-9d2941b11cb3.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.953642] env[62814]: DEBUG nova.network.neutron [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Successfully updated port: d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1065.954786] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8d12dac3-88de-4909-8bd2-0a5192cd6a51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.992704] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294369, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.996862] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1065.996862] env[62814]: value = "task-4294370" [ 1065.996862] env[62814]: _type = "Task" [ 1065.996862] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.997364] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294368, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.008245] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294370, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.130669] env[62814]: DEBUG nova.compute.manager [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1066.180458] env[62814]: DEBUG nova.network.neutron [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Updated VIF entry in instance network info cache for port dec4b0c4-68db-48fc-84e2-85becf8933bc. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1066.180834] env[62814]: DEBUG nova.network.neutron [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Updating instance_info_cache with network_info: [{"id": "dec4b0c4-68db-48fc-84e2-85becf8933bc", "address": "fa:16:3e:f5:e8:b4", "network": {"id": "afa6767f-5aa7-45a5-b700-6237ed1b36c5", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-40244317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0060ccfb18f34fbca1ab765134395bb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdec4b0c4-68", "ovs_interfaceid": "dec4b0c4-68db-48fc-84e2-85becf8933bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.301031] env[62814]: INFO nova.scheduler.client.report [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleted allocation for migration 06826230-b161-48d6-bc89-eabd48b17ffe [ 1066.372240] env[62814]: DEBUG oslo_concurrency.lockutils [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.372555] env[62814]: DEBUG oslo_concurrency.lockutils [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1066.372818] env[62814]: DEBUG nova.network.neutron [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.373101] env[62814]: DEBUG nova.objects.instance [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'info_cache' on Instance uuid 59d21ef7-df97-49ac-9329-4c18df6dd087 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.455146] env[62814]: DEBUG oslo_concurrency.lockutils [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1066.478663] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquiring lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.478663] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquired lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1066.479145] env[62814]: DEBUG nova.network.neutron [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.480021] env[62814]: DEBUG oslo_vmware.api [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294369, 'name': PowerOnVM_Task, 'duration_secs': 0.986062} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.480243] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294368, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.480635] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1066.480837] env[62814]: INFO nova.compute.manager [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Took 9.31 seconds to spawn the instance on the hypervisor. [ 1066.480837] env[62814]: DEBUG nova.compute.manager [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1066.481887] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290a32c7-8450-49f1-b03b-b5fbe82d4258 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.512134] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294370, 'name': ReconfigVM_Task, 'duration_secs': 0.462434} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.516509] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3/19b53797-6fc9-4bb0-ab3f-9d2941b11cb3.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1066.517465] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af545c55-2a7d-49a6-bf0d-9b29efaa39b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.526856] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1066.526856] env[62814]: value = "task-4294371" [ 1066.526856] env[62814]: _type = "Task" [ 1066.526856] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.545653] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294371, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.656807] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1066.683755] env[62814]: DEBUG oslo_concurrency.lockutils [req-04c4f467-9e34-488c-aff4-ba47da601965 req-6f2e68a2-9094-4f2d-875c-6992ed2f8812 service nova] Releasing lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1066.686894] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba9c8fb-78b6-48e3-a04f-f5956a7c85f2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.696194] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46c34c8e-168a-4a72-864e-7423a9c8f678 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.732283] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efeba82-8591-4c84-81c1-25495614d351 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.741371] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9c9a4d-f1da-46e0-b922-cc8840ca44b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.757919] env[62814]: DEBUG nova.compute.provider_tree [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1066.806031] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4f0f2386-f364-48d7-a8ca-ad55dfda37b5 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 25.352s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1066.945062] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "08c20588-2988-4976-8631-53cf447b4877" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1066.948017] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "08c20588-2988-4976-8631-53cf447b4877" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1066.970405] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294368, 'name': CreateVM_Task, 'duration_secs': 1.09987} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.970631] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1066.971497] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'e73b094c-8fac-409a-b8ab-24f1db58d784', 'mount_device': '/dev/sda', 'device_type': None, 'delete_on_termination': True, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845786', 'volume_id': '6f2eddd0-4334-4e22-b195-b7deaa539235', 'name': 'volume-6f2eddd0-4334-4e22-b195-b7deaa539235', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd5df1d0e-55e6-452e-882a-053a83250b39', 'attached_at': '', 'detached_at': '', 'volume_id': '6f2eddd0-4334-4e22-b195-b7deaa539235', 'serial': '6f2eddd0-4334-4e22-b195-b7deaa539235'}, 'guest_format': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=62814) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1066.971755] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Root volume attach. Driver type: vmdk {{(pid=62814) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1066.972636] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425578ea-9f83-4dd4-b7f8-40a42897ccb0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.984285] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066341c5-c44e-4e10-aeb2-a16af0c60f90 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.991450] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bad84a-f608-4412-be30-d51dc341dd36 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.003616] env[62814]: INFO nova.compute.manager [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Took 40.23 seconds to build instance. [ 1067.004536] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-0993f6d5-7858-43a9-be5d-f98a96778e12 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.013465] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for the task: (returnval){ [ 1067.013465] env[62814]: value = "task-4294372" [ 1067.013465] env[62814]: _type = "Task" [ 1067.013465] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.024302] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.027866] env[62814]: DEBUG nova.network.neutron [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1067.039739] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294371, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.176303] env[62814]: DEBUG nova.network.neutron [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updating instance_info_cache with network_info: [{"id": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "address": "fa:16:3e:3e:a0:82", "network": {"id": "6df241c7-9cf0-4d05-baba-469d1bea58ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-62005335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1c7dab0a5e874c3aaa57685a130460fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92933dd-ec", "ovs_interfaceid": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.229820] env[62814]: DEBUG nova.compute.manager [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Received event network-vif-plugged-d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1067.230610] env[62814]: DEBUG oslo_concurrency.lockutils [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] Acquiring lock "44ea319c-6ea0-456a-bee6-42133a25d8c1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1067.230610] env[62814]: DEBUG oslo_concurrency.lockutils [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] Lock "44ea319c-6ea0-456a-bee6-42133a25d8c1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1067.230610] env[62814]: DEBUG oslo_concurrency.lockutils [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] Lock "44ea319c-6ea0-456a-bee6-42133a25d8c1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.230966] env[62814]: DEBUG nova.compute.manager [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] No waiting events found dispatching network-vif-plugged-d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1067.230966] env[62814]: WARNING nova.compute.manager [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Received unexpected event network-vif-plugged-d92933dd-ecff-4c85-96cf-69211cb92fa6 for instance with vm_state building and task_state spawning. [ 1067.231072] env[62814]: DEBUG nova.compute.manager [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Received event network-changed-d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1067.231168] env[62814]: DEBUG nova.compute.manager [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Refreshing instance network info cache due to event network-changed-d92933dd-ecff-4c85-96cf-69211cb92fa6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1067.231334] env[62814]: DEBUG oslo_concurrency.lockutils [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] Acquiring lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.281989] env[62814]: ERROR nova.scheduler.client.report [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] [req-f8e4d707-74ab-4c26-9a5d-ff9b952a38a2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-f8e4d707-74ab-4c26-9a5d-ff9b952a38a2"}]} [ 1067.301084] env[62814]: DEBUG nova.scheduler.client.report [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1067.319110] env[62814]: DEBUG nova.scheduler.client.report [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1067.319375] env[62814]: DEBUG nova.compute.provider_tree [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1067.336364] env[62814]: DEBUG nova.scheduler.client.report [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1067.363016] env[62814]: DEBUG nova.scheduler.client.report [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1067.450121] env[62814]: DEBUG nova.compute.manager [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1067.508075] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ff098-b4f7-4219-b2e2-7dc7d7edb917 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "cd037f6e-fddd-4389-b6d0-144b798537bf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.329s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1067.529478] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task} progress is 40%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.545159] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294371, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.649957] env[62814]: DEBUG nova.objects.instance [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'flavor' on Instance uuid 9343301c-dfe9-41b0-b4a0-067af544d297 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.679434] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Releasing lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1067.681084] env[62814]: DEBUG nova.compute.manager [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Instance network_info: |[{"id": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "address": "fa:16:3e:3e:a0:82", "network": {"id": "6df241c7-9cf0-4d05-baba-469d1bea58ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-62005335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1c7dab0a5e874c3aaa57685a130460fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92933dd-ec", "ovs_interfaceid": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1067.681084] env[62814]: DEBUG oslo_concurrency.lockutils [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] Acquired lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1067.681084] env[62814]: DEBUG nova.network.neutron [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Refreshing network info cache for port d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.682664] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:a0:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '796c7fcb-00fd-4692-a44b-7ec550201e86', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd92933dd-ecff-4c85-96cf-69211cb92fa6', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1067.690208] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Creating folder: Project (1c7dab0a5e874c3aaa57685a130460fc). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1067.694280] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1fd8564f-a2e4-4769-9263-5621a67c183f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.707524] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Created folder: Project (1c7dab0a5e874c3aaa57685a130460fc) in parent group-v845547. [ 1067.707777] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Creating folder: Instances. Parent ref: group-v845811. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1067.710813] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6065220d-f6e5-4e9e-a67e-0434dc27566b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.723996] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Created folder: Instances in parent group-v845811. [ 1067.724824] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1067.724824] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1067.724824] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2c16d587-a2f2-4fbe-854d-8fd8c745a76d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.754300] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1067.754300] env[62814]: value = "task-4294375" [ 1067.754300] env[62814]: _type = "Task" [ 1067.754300] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.768386] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294375, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.830331] env[62814]: DEBUG nova.network.neutron [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance_info_cache with network_info: [{"id": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "address": "fa:16:3e:93:37:3e", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa00402c1-eb", "ovs_interfaceid": "a00402c1-ebc9-40c0-93b2-26b6fbab4a68", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.851916] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54241500-5381-4178-8a9e-fce3a3f4402a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.863735] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa2fcfd-101d-4ca5-b8af-45e2c37fff94 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.901086] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d822118-8b6d-4120-be47-c6d37773e8b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.908902] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eecc7e64-41c3-482f-a908-3482cd5f45e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.923779] env[62814]: DEBUG nova.compute.provider_tree [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1067.971512] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.026744] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task} progress is 53%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.043964] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294371, 'name': Rename_Task, 'duration_secs': 1.189927} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.044330] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1068.044606] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-019e9a0f-9351-4855-a1ee-85b26148bd44 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.052628] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1068.052628] env[62814]: value = "task-4294376" [ 1068.052628] env[62814]: _type = "Task" [ 1068.052628] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.061045] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294376, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.156250] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.156456] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.156702] env[62814]: DEBUG nova.network.neutron [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1068.156794] env[62814]: DEBUG nova.objects.instance [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'info_cache' on Instance uuid 9343301c-dfe9-41b0-b4a0-067af544d297 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.254506] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "cd037f6e-fddd-4389-b6d0-144b798537bf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.254910] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "cd037f6e-fddd-4389-b6d0-144b798537bf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1068.255174] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "cd037f6e-fddd-4389-b6d0-144b798537bf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1068.255405] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "cd037f6e-fddd-4389-b6d0-144b798537bf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1068.255611] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "cd037f6e-fddd-4389-b6d0-144b798537bf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1068.262330] env[62814]: INFO nova.compute.manager [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Terminating instance [ 1068.275559] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294375, 'name': CreateVM_Task, 'duration_secs': 0.428301} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.276555] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1068.277207] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.277477] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.277734] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1068.278335] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f016085-a32c-473a-ab4b-03130c1fc1b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.285657] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1068.285657] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bec1b2-6363-489e-eb36-cfe5adedabbd" [ 1068.285657] env[62814]: _type = "Task" [ 1068.285657] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.299141] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bec1b2-6363-489e-eb36-cfe5adedabbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.333954] env[62814]: DEBUG oslo_concurrency.lockutils [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-59d21ef7-df97-49ac-9329-4c18df6dd087" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.334410] env[62814]: DEBUG nova.objects.instance [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'migration_context' on Instance uuid 59d21ef7-df97-49ac-9329-4c18df6dd087 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1068.467383] env[62814]: DEBUG nova.scheduler.client.report [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 138 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1068.467786] env[62814]: DEBUG nova.compute.provider_tree [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 138 to 139 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1068.468055] env[62814]: DEBUG nova.compute.provider_tree [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 143, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1068.477391] env[62814]: DEBUG nova.network.neutron [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updated VIF entry in instance network info cache for port d92933dd-ecff-4c85-96cf-69211cb92fa6. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1068.477840] env[62814]: DEBUG nova.network.neutron [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updating instance_info_cache with network_info: [{"id": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "address": "fa:16:3e:3e:a0:82", "network": {"id": "6df241c7-9cf0-4d05-baba-469d1bea58ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-62005335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1c7dab0a5e874c3aaa57685a130460fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92933dd-ec", "ovs_interfaceid": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.532045] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task} progress is 65%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.563927] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294376, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.662495] env[62814]: DEBUG nova.objects.base [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Object Instance<9343301c-dfe9-41b0-b4a0-067af544d297> lazy-loaded attributes: flavor,info_cache {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1068.766986] env[62814]: DEBUG nova.compute.manager [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1068.767338] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1068.768262] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6735789e-b4b8-4329-8071-1909ab0ec705 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.779078] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1068.779393] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4b7cbab1-3587-4b94-97e3-d22f0c43c83d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.790220] env[62814]: DEBUG oslo_vmware.api [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1068.790220] env[62814]: value = "task-4294377" [ 1068.790220] env[62814]: _type = "Task" [ 1068.790220] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.801074] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bec1b2-6363-489e-eb36-cfe5adedabbd, 'name': SearchDatastore_Task, 'duration_secs': 0.015434} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.805681] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.805681] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1068.806191] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.806369] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1068.806640] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1068.807409] env[62814]: DEBUG oslo_vmware.api [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294377, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.807619] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a3e5c10-5c7b-4980-b0f8-85551f3edaba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.817656] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1068.817806] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1068.820180] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7db8a0a0-8992-4c57-bce5-ccb600cea8fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.824209] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1068.824209] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5235859e-4c07-2f82-e9f7-78523094efe3" [ 1068.824209] env[62814]: _type = "Task" [ 1068.824209] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.832801] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5235859e-4c07-2f82-e9f7-78523094efe3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.840891] env[62814]: DEBUG nova.objects.base [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Object Instance<59d21ef7-df97-49ac-9329-4c18df6dd087> lazy-loaded attributes: info_cache,migration_context {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1068.841971] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ac8d622-33ff-4426-96b1-b833b91d161c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.864802] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce44d2f1-ca80-4589-87b2-bd696ff0a3a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.872398] env[62814]: DEBUG oslo_vmware.api [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1068.872398] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bff389-7337-acce-6176-80bc26bfd65b" [ 1068.872398] env[62814]: _type = "Task" [ 1068.872398] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.881472] env[62814]: DEBUG oslo_vmware.api [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bff389-7337-acce-6176-80bc26bfd65b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.982400] env[62814]: DEBUG oslo_concurrency.lockutils [req-29c8ec48-09c7-402d-af82-405026efb474 req-1f20480e-9d33-4e0f-bfb7-06591b7a3ce1 service nova] Releasing lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1068.983348] env[62814]: DEBUG oslo_concurrency.lockutils [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.252s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1068.986315] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.194s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1068.986590] env[62814]: DEBUG nova.objects.instance [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lazy-loading 'resources' on Instance uuid 316931ae-7a62-4bac-81e4-1fee9a36164e {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.015310] env[62814]: INFO nova.scheduler.client.report [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Deleted allocations for instance 88ec5aba-f38d-4c75-af29-e3df3aa49640 [ 1069.031762] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task} progress is 78%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.063155] env[62814]: DEBUG oslo_vmware.api [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294376, 'name': PowerOnVM_Task, 'duration_secs': 0.704725} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.063447] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1069.063653] env[62814]: INFO nova.compute.manager [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Took 9.21 seconds to spawn the instance on the hypervisor. [ 1069.064210] env[62814]: DEBUG nova.compute.manager [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1069.065097] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c877cf1-cceb-4c70-a139-2be90eb9bbfd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.301098] env[62814]: DEBUG oslo_vmware.api [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294377, 'name': PowerOffVM_Task, 'duration_secs': 0.285881} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.301654] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1069.302011] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1069.302375] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-392f614c-d5d7-464f-92c0-c4202a68990c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.335018] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5235859e-4c07-2f82-e9f7-78523094efe3, 'name': SearchDatastore_Task, 'duration_secs': 0.025497} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.338360] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e42a655-bfe6-4b40-bcd7-31f73964d35f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.344160] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1069.344160] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528ce1b0-88ea-54a7-4fbf-8c3d8b0a4b24" [ 1069.344160] env[62814]: _type = "Task" [ 1069.344160] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.353715] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528ce1b0-88ea-54a7-4fbf-8c3d8b0a4b24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.367273] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1069.367695] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1069.368207] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleting the datastore file [datastore2] cd037f6e-fddd-4389-b6d0-144b798537bf {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.368580] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-00723725-8854-41fa-ba47-f0694592c05f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.375049] env[62814]: DEBUG oslo_vmware.api [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1069.375049] env[62814]: value = "task-4294379" [ 1069.375049] env[62814]: _type = "Task" [ 1069.375049] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.388139] env[62814]: DEBUG oslo_vmware.api [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bff389-7337-acce-6176-80bc26bfd65b, 'name': SearchDatastore_Task, 'duration_secs': 0.011261} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.391515] env[62814]: DEBUG oslo_concurrency.lockutils [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1069.392059] env[62814]: DEBUG oslo_vmware.api [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294379, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.469681] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1069.470081] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.470372] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1069.470644] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1069.471650] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.473821] env[62814]: INFO nova.compute.manager [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Terminating instance [ 1069.531866] env[62814]: DEBUG oslo_concurrency.lockutils [None req-798bb52e-3228-46b7-bde3-8d94dcd7f815 tempest-ServerAddressesTestJSON-2111633262 tempest-ServerAddressesTestJSON-2111633262-project-member] Lock "88ec5aba-f38d-4c75-af29-e3df3aa49640" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.095s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1069.542151] env[62814]: DEBUG nova.network.neutron [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance_info_cache with network_info: [{"id": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "address": "fa:16:3e:db:ee:dc", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap08bab1d2-62", "ovs_interfaceid": "08bab1d2-6296-46f7-baf6-4344d1bbb0ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.545632] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task} progress is 95%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.592030] env[62814]: INFO nova.compute.manager [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Took 41.78 seconds to build instance. [ 1069.856350] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528ce1b0-88ea-54a7-4fbf-8c3d8b0a4b24, 'name': SearchDatastore_Task, 'duration_secs': 0.022243} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.858934] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1069.859225] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 44ea319c-6ea0-456a-bee6-42133a25d8c1/44ea319c-6ea0-456a-bee6-42133a25d8c1.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1069.862193] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50c58f5e-1cb8-4911-9f10-13b1caccf92b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.867987] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1069.867987] env[62814]: value = "task-4294380" [ 1069.867987] env[62814]: _type = "Task" [ 1069.867987] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.882101] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294380, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.889835] env[62814]: DEBUG oslo_vmware.api [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294379, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.503214} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.890361] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.890592] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.890768] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.890946] env[62814]: INFO nova.compute.manager [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1069.891241] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1069.892430] env[62814]: DEBUG nova.compute.manager [-] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1069.892530] env[62814]: DEBUG nova.network.neutron [-] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1069.894650] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8c2773-f0ba-43e7-b2c3-45f74d9e41cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.901734] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666d9b60-5402-4986-bdd5-0ad56c20ffbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.941547] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fb5f87d-ef0b-4c55-b495-93e2d34d6332 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.949497] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963ecbb2-b3df-49c3-b766-eff84feb375f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.963195] env[62814]: DEBUG nova.compute.provider_tree [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1069.979876] env[62814]: DEBUG nova.compute.manager [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1069.980120] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1069.981230] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7237b0e-8633-40d1-925f-a4e11b28ff1a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.990323] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1069.990789] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f365e87-e3fd-4c88-a3de-5be3a36655ac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.997263] env[62814]: DEBUG oslo_vmware.api [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1069.997263] env[62814]: value = "task-4294381" [ 1069.997263] env[62814]: _type = "Task" [ 1069.997263] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.006589] env[62814]: DEBUG oslo_vmware.api [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294381, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.035087] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task} progress is 97%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.046601] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-9343301c-dfe9-41b0-b4a0-067af544d297" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1070.094552] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ef2832b-a2b6-41a4-96c7-4f19a2897b2a tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.314s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1070.360890] env[62814]: DEBUG nova.compute.manager [req-ae851ed6-17e7-4284-813f-f953da3cf31b req-2e6feee0-e465-4aa7-9afd-b64c7c3ce160 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Received event network-vif-deleted-317f7973-158e-46c5-89bd-1e9555150bde {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1070.360890] env[62814]: INFO nova.compute.manager [req-ae851ed6-17e7-4284-813f-f953da3cf31b req-2e6feee0-e465-4aa7-9afd-b64c7c3ce160 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Neutron deleted interface 317f7973-158e-46c5-89bd-1e9555150bde; detaching it from the instance and deleting it from the info cache [ 1070.360890] env[62814]: DEBUG nova.network.neutron [req-ae851ed6-17e7-4284-813f-f953da3cf31b req-2e6feee0-e465-4aa7-9afd-b64c7c3ce160 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.383114] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294380, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.503471] env[62814]: DEBUG nova.scheduler.client.report [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 139 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1070.503754] env[62814]: DEBUG nova.compute.provider_tree [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 139 to 140 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1070.503981] env[62814]: DEBUG nova.compute.provider_tree [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1070.513050] env[62814]: DEBUG oslo_vmware.api [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294381, 'name': PowerOffVM_Task, 'duration_secs': 0.423045} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.513548] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1070.513725] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1070.513983] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b9ddade8-fd80-4923-bab8-72eb9d76ac44 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.537774] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task} progress is 97%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.580049] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1070.580104] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1070.580328] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Deleting the datastore file [datastore2] b4ee4cee-3298-4955-8375-8ca8c04b2f9f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1070.580588] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d712eaa-57cb-46a3-92ef-dbc2fb392653 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.587706] env[62814]: DEBUG oslo_vmware.api [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1070.587706] env[62814]: value = "task-4294383" [ 1070.587706] env[62814]: _type = "Task" [ 1070.587706] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.597336] env[62814]: DEBUG oslo_vmware.api [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294383, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.797273] env[62814]: DEBUG nova.network.neutron [-] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.858834] env[62814]: INFO nova.compute.manager [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Rescuing [ 1070.858834] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.859242] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1070.859242] env[62814]: DEBUG nova.network.neutron [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1070.863740] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a1ac2103-3e45-4d4f-8197-90399e0450bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.881295] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7a3b0f-a943-4189-9bfb-12fec1659953 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.895880] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294380, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.919725] env[62814]: DEBUG nova.compute.manager [req-ae851ed6-17e7-4284-813f-f953da3cf31b req-2e6feee0-e465-4aa7-9afd-b64c7c3ce160 service nova] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Detach interface failed, port_id=317f7973-158e-46c5-89bd-1e9555150bde, reason: Instance cd037f6e-fddd-4389-b6d0-144b798537bf could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1071.015263] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.029s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1071.017854] env[62814]: DEBUG oslo_concurrency.lockutils [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.508s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1071.018260] env[62814]: DEBUG nova.objects.instance [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1071.048374] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task} progress is 97%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.053833] env[62814]: INFO nova.scheduler.client.report [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted allocations for instance 316931ae-7a62-4bac-81e4-1fee9a36164e [ 1071.057185] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1071.057628] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14924b98-2e3b-4113-86b1-c7e0cb87bf6f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.065906] env[62814]: DEBUG oslo_vmware.api [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1071.065906] env[62814]: value = "task-4294384" [ 1071.065906] env[62814]: _type = "Task" [ 1071.065906] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.083821] env[62814]: DEBUG oslo_vmware.api [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294384, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.097291] env[62814]: DEBUG oslo_vmware.api [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294383, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.300202] env[62814]: INFO nova.compute.manager [-] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Took 1.41 seconds to deallocate network for instance. [ 1071.384818] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294380, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.10185} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.384818] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 44ea319c-6ea0-456a-bee6-42133a25d8c1/44ea319c-6ea0-456a-bee6-42133a25d8c1.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1071.384818] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1071.384818] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e38820eb-b557-4241-8566-dc7452f893c7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.390992] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1071.390992] env[62814]: value = "task-4294385" [ 1071.390992] env[62814]: _type = "Task" [ 1071.390992] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.402356] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294385, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.541256] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task} progress is 98%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.565094] env[62814]: DEBUG oslo_concurrency.lockutils [None req-97f51d7d-47af-48a3-a58a-1a69abdefd44 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "316931ae-7a62-4bac-81e4-1fee9a36164e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.148s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1071.578972] env[62814]: DEBUG oslo_vmware.api [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294384, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.597801] env[62814]: DEBUG oslo_vmware.api [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294383, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.683747} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.598828] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1071.599064] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1071.599300] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1071.599570] env[62814]: INFO nova.compute.manager [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1071.600021] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1071.601021] env[62814]: DEBUG nova.compute.manager [-] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1071.601021] env[62814]: DEBUG nova.network.neutron [-] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1071.708469] env[62814]: DEBUG nova.network.neutron [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Updating instance_info_cache with network_info: [{"id": "2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f", "address": "fa:16:3e:b5:d9:f0", "network": {"id": "dbc702b4-6191-4068-a1d6-4c03cc349962", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1378843546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d8b042739d3b4927a28b9729ec5dc08b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2abec6f6-ac", "ovs_interfaceid": "2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1071.810379] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1071.903013] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294385, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067431} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.903013] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1071.903628] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b6653d-180f-46ce-8d76-d4044c63dbae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.936713] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 44ea319c-6ea0-456a-bee6-42133a25d8c1/44ea319c-6ea0-456a-bee6-42133a25d8c1.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.937165] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c0d6ff27-eb9e-4ca9-a2b4-4c380b30ebff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.966015] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1071.966015] env[62814]: value = "task-4294386" [ 1071.966015] env[62814]: _type = "Task" [ 1071.966015] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.977753] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294386, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.026212] env[62814]: DEBUG nova.compute.manager [req-69d04226-6b5f-4e3c-9355-497f702e52e1 req-06c4bb67-7f83-4c4f-81f7-c42534209fb1 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Received event network-vif-deleted-89bae162-2360-4722-8b1c-ade210004fcc {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1072.026421] env[62814]: INFO nova.compute.manager [req-69d04226-6b5f-4e3c-9355-497f702e52e1 req-06c4bb67-7f83-4c4f-81f7-c42534209fb1 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Neutron deleted interface 89bae162-2360-4722-8b1c-ade210004fcc; detaching it from the instance and deleting it from the info cache [ 1072.026595] env[62814]: DEBUG nova.network.neutron [req-69d04226-6b5f-4e3c-9355-497f702e52e1 req-06c4bb67-7f83-4c4f-81f7-c42534209fb1 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.030551] env[62814]: DEBUG oslo_concurrency.lockutils [None req-162fe93a-acd7-4137-8052-a769deecd5a3 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1072.033563] env[62814]: DEBUG oslo_concurrency.lockutils [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.668s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1072.033791] env[62814]: DEBUG nova.objects.instance [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lazy-loading 'resources' on Instance uuid c7cef7f3-11db-44e1-a454-98830b465b52 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.049384] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294372, 'name': RelocateVM_Task, 'duration_secs': 4.846311} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.049728] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1072.049965] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845786', 'volume_id': '6f2eddd0-4334-4e22-b195-b7deaa539235', 'name': 'volume-6f2eddd0-4334-4e22-b195-b7deaa539235', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd5df1d0e-55e6-452e-882a-053a83250b39', 'attached_at': '', 'detached_at': '', 'volume_id': '6f2eddd0-4334-4e22-b195-b7deaa539235', 'serial': '6f2eddd0-4334-4e22-b195-b7deaa539235'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1072.051867] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d085f8-ae7e-4d20-a0a2-b9f282ea615a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.073558] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8874943f-11ab-453b-8c24-d47df7eab724 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.081432] env[62814]: DEBUG oslo_vmware.api [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294384, 'name': PowerOnVM_Task, 'duration_secs': 0.540066} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.090694] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1072.090975] env[62814]: DEBUG nova.compute.manager [None req-4c65652d-48d0-458e-98d9-8864babe5954 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1072.101180] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] volume-6f2eddd0-4334-4e22-b195-b7deaa539235/volume-6f2eddd0-4334-4e22-b195-b7deaa539235.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1072.102124] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1036911-ccda-464f-9a50-90fc0fc8c635 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.106729] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a07808bd-4f1d-44e0-8bec-de9b30c90b34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.127587] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for the task: (returnval){ [ 1072.127587] env[62814]: value = "task-4294387" [ 1072.127587] env[62814]: _type = "Task" [ 1072.127587] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.136823] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294387, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.211599] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1072.461834] env[62814]: DEBUG nova.network.neutron [-] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.479978] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294386, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.535257] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc7f2fc4-af7c-4c1f-9718-9d37a1e19a9b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.549765] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a4b72ce-7366-4675-a28b-9258f937e410 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.587016] env[62814]: DEBUG nova.compute.manager [req-69d04226-6b5f-4e3c-9355-497f702e52e1 req-06c4bb67-7f83-4c4f-81f7-c42534209fb1 service nova] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Detach interface failed, port_id=89bae162-2360-4722-8b1c-ade210004fcc, reason: Instance b4ee4cee-3298-4955-8375-8ca8c04b2f9f could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1072.641781] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294387, 'name': ReconfigVM_Task, 'duration_secs': 0.28146} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.642078] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Reconfigured VM instance instance-0000005b to attach disk [datastore2] volume-6f2eddd0-4334-4e22-b195-b7deaa539235/volume-6f2eddd0-4334-4e22-b195-b7deaa539235.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1072.647081] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d1f02ce-d06d-4976-bde7-0d97fede77f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.663630] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for the task: (returnval){ [ 1072.663630] env[62814]: value = "task-4294388" [ 1072.663630] env[62814]: _type = "Task" [ 1072.663630] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.674421] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294388, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.965382] env[62814]: INFO nova.compute.manager [-] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Took 1.36 seconds to deallocate network for instance. [ 1072.971779] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4ad050-8712-4469-971d-651bcdf5f21d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.987588] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294386, 'name': ReconfigVM_Task, 'duration_secs': 0.576248} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.994175] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 44ea319c-6ea0-456a-bee6-42133a25d8c1/44ea319c-6ea0-456a-bee6-42133a25d8c1.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1072.994175] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4844beb-2823-471f-b4ea-f486fbad5dad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.994874] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9998929d-5ac4-414c-a55f-dca7f1902be7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.043669] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b889679-083b-4817-931a-94a5344ff233 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.046645] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1073.046645] env[62814]: value = "task-4294389" [ 1073.046645] env[62814]: _type = "Task" [ 1073.046645] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.055703] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7961bc74-cd86-4d56-a1cd-abf9f5456225 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.063730] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294389, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.075219] env[62814]: DEBUG nova.compute.provider_tree [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1073.185362] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294388, 'name': ReconfigVM_Task, 'duration_secs': 0.126041} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.185571] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845786', 'volume_id': '6f2eddd0-4334-4e22-b195-b7deaa539235', 'name': 'volume-6f2eddd0-4334-4e22-b195-b7deaa539235', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd5df1d0e-55e6-452e-882a-053a83250b39', 'attached_at': '', 'detached_at': '', 'volume_id': '6f2eddd0-4334-4e22-b195-b7deaa539235', 'serial': '6f2eddd0-4334-4e22-b195-b7deaa539235'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1073.186257] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f8408dd-8525-438b-a3ed-29d6f64b2ddd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.195454] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for the task: (returnval){ [ 1073.195454] env[62814]: value = "task-4294390" [ 1073.195454] env[62814]: _type = "Task" [ 1073.195454] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.209176] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294390, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.480426] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1073.557846] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294389, 'name': Rename_Task, 'duration_secs': 0.14709} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.560018] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1073.560018] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf7cdc04-bf2a-4a6f-932c-6ac06f7c6093 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.565754] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1073.565754] env[62814]: value = "task-4294391" [ 1073.565754] env[62814]: _type = "Task" [ 1073.565754] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.574398] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294391, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.579696] env[62814]: DEBUG nova.scheduler.client.report [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1073.706576] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294390, 'name': Rename_Task, 'duration_secs': 0.253368} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.706988] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1073.707287] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5713009e-e811-4234-921f-a825a59e43d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.714747] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for the task: (returnval){ [ 1073.714747] env[62814]: value = "task-4294392" [ 1073.714747] env[62814]: _type = "Task" [ 1073.714747] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.727889] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.755752] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1073.757174] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-530aebc9-6022-4909-af0a-17d8a78ed637 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.763779] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1073.763779] env[62814]: value = "task-4294393" [ 1073.763779] env[62814]: _type = "Task" [ 1073.763779] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.776727] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.790394] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "8af98433-18eb-4a12-b18d-39b0a58821c0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1073.790636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.080869] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294391, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.084995] env[62814]: DEBUG oslo_concurrency.lockutils [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.051s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.090020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.723s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.091965] env[62814]: INFO nova.compute.claims [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1074.122781] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "9343301c-dfe9-41b0-b4a0-067af544d297" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.122781] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.123035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "9343301c-dfe9-41b0-b4a0-067af544d297-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.123741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1074.123741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.126530] env[62814]: INFO nova.compute.manager [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Terminating instance [ 1074.145296] env[62814]: INFO nova.scheduler.client.report [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Deleted allocations for instance c7cef7f3-11db-44e1-a454-98830b465b52 [ 1074.224894] env[62814]: DEBUG oslo_vmware.api [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294392, 'name': PowerOnVM_Task, 'duration_secs': 0.496292} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.225260] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1074.225482] env[62814]: INFO nova.compute.manager [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Took 9.84 seconds to spawn the instance on the hypervisor. [ 1074.225663] env[62814]: DEBUG nova.compute.manager [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1074.226486] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9287daf-15e0-4d3b-8de0-7a9ba97cf6a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.273750] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294393, 'name': PowerOffVM_Task, 'duration_secs': 0.241551} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.274950] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1074.275977] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64b44e16-b6e4-4e97-a8dc-c46a98eee9e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.298071] env[62814]: DEBUG nova.compute.manager [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1074.304071] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8393465f-e00e-4aa5-9673-3a515ef3748b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.342731] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1074.343077] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b0ff54d-7617-4f22-9a1f-883c823dc203 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.350715] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1074.350715] env[62814]: value = "task-4294394" [ 1074.350715] env[62814]: _type = "Task" [ 1074.350715] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.361836] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1074.362101] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1074.362351] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1074.362499] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1074.362673] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1074.362951] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab62c197-1c49-4b42-9894-e2d5d808a09a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.373356] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1074.373510] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1074.374397] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5438ef88-38f8-4dec-9afb-e6eb2c592745 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.379377] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1074.379377] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527e1714-9fc5-e12d-b524-480b125d0f6d" [ 1074.379377] env[62814]: _type = "Task" [ 1074.379377] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.388110] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527e1714-9fc5-e12d-b524-480b125d0f6d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.583487] env[62814]: DEBUG oslo_vmware.api [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294391, 'name': PowerOnVM_Task, 'duration_secs': 0.556382} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.583939] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1074.584374] env[62814]: INFO nova.compute.manager [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Took 9.39 seconds to spawn the instance on the hypervisor. [ 1074.584624] env[62814]: DEBUG nova.compute.manager [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1074.586748] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52d22d4-23f0-4365-9b34-e78818bb2d58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.632546] env[62814]: DEBUG nova.compute.manager [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1074.632772] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1074.634128] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a650a59-349b-489b-9a08-6d9dac9b8cc2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.641640] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1074.642308] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c955a1e-6c95-44d2-aea0-d57c390885e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.649236] env[62814]: DEBUG oslo_vmware.api [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1074.649236] env[62814]: value = "task-4294395" [ 1074.649236] env[62814]: _type = "Task" [ 1074.649236] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.656390] env[62814]: DEBUG oslo_concurrency.lockutils [None req-60d32e4f-3492-4aa1-9f45-414d3407113c tempest-ServerRescueNegativeTestJSON-253242598 tempest-ServerRescueNegativeTestJSON-253242598-project-member] Lock "c7cef7f3-11db-44e1-a454-98830b465b52" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.038s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1074.664772] env[62814]: DEBUG oslo_vmware.api [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294395, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.750096] env[62814]: INFO nova.compute.manager [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Took 40.67 seconds to build instance. [ 1074.821443] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1074.894013] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527e1714-9fc5-e12d-b524-480b125d0f6d, 'name': SearchDatastore_Task, 'duration_secs': 0.028848} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.904033] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2082ebf-b0b7-4ebf-9509-892e4e9b82a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.907371] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1074.907371] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52354ac4-feb3-d492-a3c7-c3e19f851a2e" [ 1074.907371] env[62814]: _type = "Task" [ 1074.907371] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.918013] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52354ac4-feb3-d492-a3c7-c3e19f851a2e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.116667] env[62814]: INFO nova.compute.manager [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Took 38.79 seconds to build instance. [ 1075.163457] env[62814]: DEBUG oslo_vmware.api [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294395, 'name': PowerOffVM_Task, 'duration_secs': 0.193252} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.163902] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1075.164221] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1075.164593] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6741964f-1919-4091-b2e7-f29409755d8b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.239160] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1075.239382] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1075.239563] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleting the datastore file [datastore2] 9343301c-dfe9-41b0-b4a0-067af544d297 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1075.240379] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4b262c64-66c0-4cb5-a02b-438993ca8cad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.249543] env[62814]: DEBUG oslo_vmware.api [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1075.249543] env[62814]: value = "task-4294397" [ 1075.249543] env[62814]: _type = "Task" [ 1075.249543] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.254103] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b36dc794-7034-430e-bbc8-4b42fa2b7ab7 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "d5df1d0e-55e6-452e-882a-053a83250b39" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.079s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.259497] env[62814]: DEBUG oslo_vmware.api [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294397, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.427195] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52354ac4-feb3-d492-a3c7-c3e19f851a2e, 'name': SearchDatastore_Task, 'duration_secs': 0.009076} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.427547] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1075.427872] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. {{(pid=62814) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1075.428339] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5941f368-2b00-4299-97ea-544c86ac2ea7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.437863] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1075.437863] env[62814]: value = "task-4294398" [ 1075.437863] env[62814]: _type = "Task" [ 1075.437863] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.445331] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294398, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.560020] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9e2688f-79e5-4500-b813-39c57f9d1038 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.567351] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96a27cc9-781b-4677-b7a5-2af1e4720d44 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.600548] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94179eb1-8152-4ed0-bd78-5ada15b98ec6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.608829] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7af68be3-3d29-40cc-adb7-60d9fd1842cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.623944] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9d911570-4aac-45e9-b799-79363c615d32 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lock "44ea319c-6ea0-456a-bee6-42133a25d8c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.910s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1075.624532] env[62814]: DEBUG nova.compute.provider_tree [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1075.769490] env[62814]: DEBUG oslo_vmware.api [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294397, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.279101} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.769864] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1075.772133] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1075.772133] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1075.772133] env[62814]: INFO nova.compute.manager [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1075.772133] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1075.772133] env[62814]: DEBUG nova.compute.manager [-] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1075.772133] env[62814]: DEBUG nova.network.neutron [-] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1075.952475] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294398, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.127930] env[62814]: DEBUG nova.scheduler.client.report [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1076.456914] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294398, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568438} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.457281] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. [ 1076.458109] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e99ef4-b1c2-492a-9241-f0b460c40066 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.489407] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1076.489836] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-adc71555-9d9d-4cb4-9ca6-44b6f74b9212 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.512827] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1076.512827] env[62814]: value = "task-4294399" [ 1076.512827] env[62814]: _type = "Task" [ 1076.512827] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.526022] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294399, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.561231] env[62814]: INFO nova.compute.manager [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Rescuing [ 1076.561573] env[62814]: DEBUG oslo_concurrency.lockutils [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquiring lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.561797] env[62814]: DEBUG oslo_concurrency.lockutils [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquired lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1076.562016] env[62814]: DEBUG nova.network.neutron [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1076.635036] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1076.635613] env[62814]: DEBUG nova.compute.manager [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1076.641665] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.862s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1076.642013] env[62814]: DEBUG nova.objects.instance [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lazy-loading 'resources' on Instance uuid f54b8bb9-69b7-4bb4-a82c-9f796050e719 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.712202] env[62814]: DEBUG nova.compute.manager [req-5d468870-dc01-4df7-a99e-e67a43a846b1 req-4bd58846-92d7-4b18-bc26-848aa1737b9b service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Received event network-changed-dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1076.712202] env[62814]: DEBUG nova.compute.manager [req-5d468870-dc01-4df7-a99e-e67a43a846b1 req-4bd58846-92d7-4b18-bc26-848aa1737b9b service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Refreshing instance network info cache due to event network-changed-dec4b0c4-68db-48fc-84e2-85becf8933bc. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1076.712992] env[62814]: DEBUG oslo_concurrency.lockutils [req-5d468870-dc01-4df7-a99e-e67a43a846b1 req-4bd58846-92d7-4b18-bc26-848aa1737b9b service nova] Acquiring lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.713229] env[62814]: DEBUG oslo_concurrency.lockutils [req-5d468870-dc01-4df7-a99e-e67a43a846b1 req-4bd58846-92d7-4b18-bc26-848aa1737b9b service nova] Acquired lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1076.713463] env[62814]: DEBUG nova.network.neutron [req-5d468870-dc01-4df7-a99e-e67a43a846b1 req-4bd58846-92d7-4b18-bc26-848aa1737b9b service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Refreshing network info cache for port dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1077.024745] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294399, 'name': ReconfigVM_Task, 'duration_secs': 0.302941} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.024853] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1077.025714] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3fecc2-1af0-4ac1-8a1b-989ea2466eb0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.052962] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5af18dd-e412-4a11-b4f3-b94741c02a1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.070122] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1077.070122] env[62814]: value = "task-4294400" [ 1077.070122] env[62814]: _type = "Task" [ 1077.070122] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.081095] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294400, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.150716] env[62814]: DEBUG nova.compute.utils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1077.155146] env[62814]: DEBUG nova.compute.manager [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1077.155255] env[62814]: DEBUG nova.network.neutron [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1077.322657] env[62814]: DEBUG nova.policy [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64e84da705284d2dbf693a26ef184cd7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95c92336f9e746edba50b0b9e078b0dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1077.330509] env[62814]: DEBUG nova.network.neutron [-] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.457254] env[62814]: DEBUG nova.network.neutron [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updating instance_info_cache with network_info: [{"id": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "address": "fa:16:3e:3e:a0:82", "network": {"id": "6df241c7-9cf0-4d05-baba-469d1bea58ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-62005335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1c7dab0a5e874c3aaa57685a130460fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92933dd-ec", "ovs_interfaceid": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.587964] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294400, 'name': ReconfigVM_Task, 'duration_secs': 0.168349} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.588519] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1077.588781] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d218f2fb-49be-4d87-a5e5-f8611c61846e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.599767] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1077.599767] env[62814]: value = "task-4294401" [ 1077.599767] env[62814]: _type = "Task" [ 1077.599767] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.611420] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294401, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.660497] env[62814]: DEBUG nova.compute.manager [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1077.683869] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05862aa1-4a23-4b85-ab20-40df945022d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.695023] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653f388f-e4b3-40da-b4a6-9d5a8e1f9703 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.734906] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af02496-7c39-4383-8cb0-1bfe98cbb962 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.744404] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe355879-c089-4d48-9edb-61c40b9a4353 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.761582] env[62814]: DEBUG nova.compute.provider_tree [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.833503] env[62814]: INFO nova.compute.manager [-] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Took 2.06 seconds to deallocate network for instance. [ 1077.926160] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "1f13051e-e51b-4981-9445-d5420c3c1818" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1077.926539] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1077.960191] env[62814]: DEBUG oslo_concurrency.lockutils [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Releasing lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.035492] env[62814]: DEBUG nova.network.neutron [req-5d468870-dc01-4df7-a99e-e67a43a846b1 req-4bd58846-92d7-4b18-bc26-848aa1737b9b service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Updated VIF entry in instance network info cache for port dec4b0c4-68db-48fc-84e2-85becf8933bc. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1078.035854] env[62814]: DEBUG nova.network.neutron [req-5d468870-dc01-4df7-a99e-e67a43a846b1 req-4bd58846-92d7-4b18-bc26-848aa1737b9b service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Updating instance_info_cache with network_info: [{"id": "dec4b0c4-68db-48fc-84e2-85becf8933bc", "address": "fa:16:3e:f5:e8:b4", "network": {"id": "afa6767f-5aa7-45a5-b700-6237ed1b36c5", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-40244317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0060ccfb18f34fbca1ab765134395bb8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e350f83a-f581-4e10-ac16-0b0f7bfd3d38", "external-id": "nsx-vlan-transportzone-834", "segmentation_id": 834, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdec4b0c4-68", "ovs_interfaceid": "dec4b0c4-68db-48fc-84e2-85becf8933bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.114512] env[62814]: DEBUG oslo_vmware.api [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294401, 'name': PowerOnVM_Task, 'duration_secs': 0.447811} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.114814] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1078.117607] env[62814]: DEBUG nova.compute.manager [None req-b02f9329-b5e6-4a38-8bb6-72df9c2abfde tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1078.118408] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6d5a8a-a442-4e5d-ae5d-6411efd4560c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.212263] env[62814]: DEBUG nova.network.neutron [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Successfully created port: f1881a6c-5678-4b14-9eb9-f3e7f3e41810 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1078.267214] env[62814]: DEBUG nova.scheduler.client.report [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1078.343777] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1078.428458] env[62814]: DEBUG nova.compute.manager [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1078.538688] env[62814]: DEBUG oslo_concurrency.lockutils [req-5d468870-dc01-4df7-a99e-e67a43a846b1 req-4bd58846-92d7-4b18-bc26-848aa1737b9b service nova] Releasing lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1078.674278] env[62814]: DEBUG nova.compute.manager [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1078.703959] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1078.704254] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1078.704423] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1078.704606] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1078.704752] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1078.704899] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1078.705187] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1078.705288] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1078.705455] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1078.705616] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1078.705791] env[62814]: DEBUG nova.virt.hardware [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1078.706697] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c71edc2-e85b-4f5b-a5a5-2a3f56ffc452 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.715445] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7284c9fc-7bc3-439b-9a35-5ad8aa82922c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.774317] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.133s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1078.777421] env[62814]: DEBUG nova.compute.manager [req-e5768095-4fe9-4279-bc5c-9bbed8c7d081 req-7129719a-1dda-49e8-adb2-4887128a1824 service nova] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Received event network-vif-deleted-08bab1d2-6296-46f7-baf6-4344d1bbb0ef {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1078.778259] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.371s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1078.778259] env[62814]: DEBUG nova.objects.instance [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lazy-loading 'resources' on Instance uuid a07b964f-fe88-4c88-b6f4-61ed0973716c {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.803417] env[62814]: INFO nova.scheduler.client.report [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Deleted allocations for instance f54b8bb9-69b7-4bb4-a82c-9f796050e719 [ 1078.964269] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1079.312347] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4200aeb0-4cb4-4669-bfee-00e0f1b28fef tempest-ListImageFiltersTestJSON-1926878691 tempest-ListImageFiltersTestJSON-1926878691-project-member] Lock "f54b8bb9-69b7-4bb4-a82c-9f796050e719" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.136s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.508183] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1079.508589] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d311d4ef-7d33-4e11-81fe-82897a5d342c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.519537] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1079.519537] env[62814]: value = "task-4294402" [ 1079.519537] env[62814]: _type = "Task" [ 1079.519537] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.530562] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294402, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.614085] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8a1626-3860-4a64-8c17-8a30aee0045e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.624387] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d86642-719e-462d-8f10-ea76d2684547 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.660487] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edcf01a7-9e4f-4821-9a31-db63147fb164 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.663290] env[62814]: INFO nova.compute.manager [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Unrescuing [ 1079.663534] env[62814]: DEBUG oslo_concurrency.lockutils [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1079.663682] env[62814]: DEBUG oslo_concurrency.lockutils [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquired lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1079.663844] env[62814]: DEBUG nova.network.neutron [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1079.671095] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ace011d-01de-4be5-9265-05cd5e79f870 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.685690] env[62814]: DEBUG nova.compute.provider_tree [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.981649] env[62814]: DEBUG nova.compute.manager [req-9ed3f8a7-0bb0-4ac9-8201-cadf2a323522 req-630f805f-3454-456d-892d-6a43aaa47beb service nova] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Received event network-vif-plugged-f1881a6c-5678-4b14-9eb9-f3e7f3e41810 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1079.982630] env[62814]: DEBUG oslo_concurrency.lockutils [req-9ed3f8a7-0bb0-4ac9-8201-cadf2a323522 req-630f805f-3454-456d-892d-6a43aaa47beb service nova] Acquiring lock "7e849b23-226d-4a1d-a85c-cf96964ce034-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1079.982931] env[62814]: DEBUG oslo_concurrency.lockutils [req-9ed3f8a7-0bb0-4ac9-8201-cadf2a323522 req-630f805f-3454-456d-892d-6a43aaa47beb service nova] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1079.984850] env[62814]: DEBUG oslo_concurrency.lockutils [req-9ed3f8a7-0bb0-4ac9-8201-cadf2a323522 req-630f805f-3454-456d-892d-6a43aaa47beb service nova] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1079.984850] env[62814]: DEBUG nova.compute.manager [req-9ed3f8a7-0bb0-4ac9-8201-cadf2a323522 req-630f805f-3454-456d-892d-6a43aaa47beb service nova] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] No waiting events found dispatching network-vif-plugged-f1881a6c-5678-4b14-9eb9-f3e7f3e41810 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1079.984850] env[62814]: WARNING nova.compute.manager [req-9ed3f8a7-0bb0-4ac9-8201-cadf2a323522 req-630f805f-3454-456d-892d-6a43aaa47beb service nova] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Received unexpected event network-vif-plugged-f1881a6c-5678-4b14-9eb9-f3e7f3e41810 for instance with vm_state building and task_state spawning. [ 1080.029857] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294402, 'name': PowerOffVM_Task, 'duration_secs': 0.318736} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.030567] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1080.030984] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f91b75-462c-4095-9250-45101fdfa91e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.050132] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c6bdb5a-8a8f-479b-82a4-e3f56cbd990d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.084206] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.084570] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f45cb2c-add6-4334-87a8-1ff49dc1c985 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.092263] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1080.092263] env[62814]: value = "task-4294403" [ 1080.092263] env[62814]: _type = "Task" [ 1080.092263] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.103498] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294403, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.110024] env[62814]: DEBUG nova.network.neutron [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Successfully updated port: f1881a6c-5678-4b14-9eb9-f3e7f3e41810 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1080.188908] env[62814]: DEBUG nova.scheduler.client.report [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1080.439843] env[62814]: DEBUG nova.network.neutron [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Updating instance_info_cache with network_info: [{"id": "2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f", "address": "fa:16:3e:b5:d9:f0", "network": {"id": "dbc702b4-6191-4068-a1d6-4c03cc349962", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1378843546-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.2", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "d8b042739d3b4927a28b9729ec5dc08b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e6f11c0d-c73a-47f5-b02e-47bff48da0e4", "external-id": "nsx-vlan-transportzone-345", "segmentation_id": 345, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2abec6f6-ac", "ovs_interfaceid": "2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.606027] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1080.606027] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1080.606027] env[62814]: DEBUG oslo_concurrency.lockutils [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.606027] env[62814]: DEBUG oslo_concurrency.lockutils [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1080.606027] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1080.606027] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b603102-16e4-4717-bff0-f060f81d7991 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.610961] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.610961] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1080.611162] env[62814]: DEBUG nova.network.neutron [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1080.615378] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1080.615579] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1080.616336] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da002832-d647-444a-807b-c1f00ca85c05 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.624341] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1080.624341] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aade31-8466-232e-8f98-496f2a22cfd8" [ 1080.624341] env[62814]: _type = "Task" [ 1080.624341] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.635685] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aade31-8466-232e-8f98-496f2a22cfd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.697971] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.920s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1080.700715] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.077s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1080.701066] env[62814]: DEBUG nova.objects.instance [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lazy-loading 'resources' on Instance uuid 1e259ec6-d31a-453b-87e0-baa446665d56 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.725654] env[62814]: INFO nova.scheduler.client.report [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted allocations for instance a07b964f-fe88-4c88-b6f4-61ed0973716c [ 1080.943867] env[62814]: DEBUG oslo_concurrency.lockutils [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Releasing lock "refresh_cache-19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1080.944521] env[62814]: DEBUG nova.objects.instance [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lazy-loading 'flavor' on Instance uuid 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1081.137963] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aade31-8466-232e-8f98-496f2a22cfd8, 'name': SearchDatastore_Task, 'duration_secs': 0.02343} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.139222] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd42dfbe-f277-4ada-bed1-f0902418ebac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.145185] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1081.145185] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5273a640-23f1-3576-c92d-23a856307775" [ 1081.145185] env[62814]: _type = "Task" [ 1081.145185] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.154705] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5273a640-23f1-3576-c92d-23a856307775, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.158245] env[62814]: DEBUG nova.network.neutron [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1081.235561] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cddf76e4-f70d-42ea-b8d0-acb855ecffce tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "a07b964f-fe88-4c88-b6f4-61ed0973716c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.802s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.363269] env[62814]: DEBUG nova.network.neutron [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance_info_cache with network_info: [{"id": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "address": "fa:16:3e:35:63:2b", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1881a6c-56", "ovs_interfaceid": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.426682] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "9683894b-a300-4400-a1b9-db62478f42c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.427561] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.428248] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "9683894b-a300-4400-a1b9-db62478f42c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1081.428461] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1081.430201] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1081.431840] env[62814]: INFO nova.compute.manager [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Terminating instance [ 1081.452560] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79794751-f560-4350-a77f-056ef4808fbd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.477744] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1081.480665] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3848499a-d3cd-43b0-bbfb-3f1b559be441 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.489387] env[62814]: DEBUG oslo_vmware.api [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1081.489387] env[62814]: value = "task-4294404" [ 1081.489387] env[62814]: _type = "Task" [ 1081.489387] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.501768] env[62814]: DEBUG oslo_vmware.api [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294404, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.628372] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4444352d-03fd-41e4-bac9-78701ba9a2e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.636888] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2afc65b-cca8-4695-a18f-cc495a711358 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.671382] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61827812-ead8-4e41-aab9-87e07599e4e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.682904] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12fee51d-17f1-4fe2-b3dc-6207a0ae93b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.687663] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5273a640-23f1-3576-c92d-23a856307775, 'name': SearchDatastore_Task, 'duration_secs': 0.031275} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.687839] env[62814]: DEBUG oslo_concurrency.lockutils [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1081.688484] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 44ea319c-6ea0-456a-bee6-42133a25d8c1/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. {{(pid=62814) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1081.689154] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2289a6b9-2fda-4b1b-bcc8-944500806608 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.699289] env[62814]: DEBUG nova.compute.provider_tree [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1081.708797] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1081.708797] env[62814]: value = "task-4294405" [ 1081.708797] env[62814]: _type = "Task" [ 1081.708797] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.718816] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294405, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.872172] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1081.872172] env[62814]: DEBUG nova.compute.manager [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Instance network_info: |[{"id": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "address": "fa:16:3e:35:63:2b", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1881a6c-56", "ovs_interfaceid": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1081.872172] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:63:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '605f83bd-808c-4b54-922e-54b14690987a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f1881a6c-5678-4b14-9eb9-f3e7f3e41810', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1081.882260] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1081.883105] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1081.883549] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba93dbec-efd0-43be-abf1-db32b548c36c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.910978] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1081.910978] env[62814]: value = "task-4294406" [ 1081.910978] env[62814]: _type = "Task" [ 1081.910978] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.920207] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294406, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.939317] env[62814]: DEBUG nova.compute.manager [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1081.939540] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1081.941576] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40277cc1-3ef6-4cd3-9925-6021145e46af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.957138] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1081.957453] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d377b89-5239-4047-a052-0cd5eeaba567 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.965896] env[62814]: DEBUG oslo_vmware.api [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1081.965896] env[62814]: value = "task-4294407" [ 1081.965896] env[62814]: _type = "Task" [ 1081.965896] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.975235] env[62814]: DEBUG oslo_vmware.api [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294407, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.001200] env[62814]: DEBUG oslo_vmware.api [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294404, 'name': PowerOffVM_Task, 'duration_secs': 0.286766} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.002512] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1082.009053] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Reconfiguring VM instance instance-0000005a to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1082.009514] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65ae1278-2ea4-40a6-9592-a2a26bce347b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.034134] env[62814]: DEBUG oslo_vmware.api [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1082.034134] env[62814]: value = "task-4294408" [ 1082.034134] env[62814]: _type = "Task" [ 1082.034134] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.048027] env[62814]: DEBUG oslo_vmware.api [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294408, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.066484] env[62814]: DEBUG nova.compute.manager [req-3083047d-4ede-44b2-a5a8-5af77eca4af2 req-161fc7af-9b7f-410a-a79c-458fdf3cfde1 service nova] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Received event network-changed-f1881a6c-5678-4b14-9eb9-f3e7f3e41810 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1082.066818] env[62814]: DEBUG nova.compute.manager [req-3083047d-4ede-44b2-a5a8-5af77eca4af2 req-161fc7af-9b7f-410a-a79c-458fdf3cfde1 service nova] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Refreshing instance network info cache due to event network-changed-f1881a6c-5678-4b14-9eb9-f3e7f3e41810. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1082.067799] env[62814]: DEBUG oslo_concurrency.lockutils [req-3083047d-4ede-44b2-a5a8-5af77eca4af2 req-161fc7af-9b7f-410a-a79c-458fdf3cfde1 service nova] Acquiring lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.068029] env[62814]: DEBUG oslo_concurrency.lockutils [req-3083047d-4ede-44b2-a5a8-5af77eca4af2 req-161fc7af-9b7f-410a-a79c-458fdf3cfde1 service nova] Acquired lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.068254] env[62814]: DEBUG nova.network.neutron [req-3083047d-4ede-44b2-a5a8-5af77eca4af2 req-161fc7af-9b7f-410a-a79c-458fdf3cfde1 service nova] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Refreshing network info cache for port f1881a6c-5678-4b14-9eb9-f3e7f3e41810 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1082.207636] env[62814]: DEBUG nova.scheduler.client.report [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1082.223842] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294405, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.421510] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294406, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.477326] env[62814]: DEBUG oslo_vmware.api [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294407, 'name': PowerOffVM_Task, 'duration_secs': 0.30944} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.478321] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1082.478722] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1082.479096] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8e695554-00a1-45be-b631-0169f7194e89 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.549313] env[62814]: DEBUG oslo_vmware.api [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294408, 'name': ReconfigVM_Task, 'duration_secs': 0.469986} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.550687] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Reconfigured VM instance instance-0000005a to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1082.550953] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.551160] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1082.551340] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1082.551510] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Deleting the datastore file [datastore2] 9683894b-a300-4400-a1b9-db62478f42c5 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1082.551744] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc17786f-5e98-43b1-88fa-2af23cf8029c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.553503] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42173edb-6c72-45d1-aad5-2d3df4b6e35d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.559089] env[62814]: DEBUG oslo_vmware.api [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1082.559089] env[62814]: value = "task-4294410" [ 1082.559089] env[62814]: _type = "Task" [ 1082.559089] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.560561] env[62814]: DEBUG oslo_vmware.api [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1082.560561] env[62814]: value = "task-4294411" [ 1082.560561] env[62814]: _type = "Task" [ 1082.560561] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.574071] env[62814]: DEBUG oslo_vmware.api [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294410, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.578842] env[62814]: DEBUG oslo_vmware.api [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294411, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.714523] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.014s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1082.720028] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.889s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1082.721756] env[62814]: INFO nova.compute.claims [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1082.734610] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294405, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5232} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.734862] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 44ea319c-6ea0-456a-bee6-42133a25d8c1/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk. [ 1082.735699] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21fa7cea-4b24-470f-a3e6-beb44007138f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.765264] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 44ea319c-6ea0-456a-bee6-42133a25d8c1/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.766568] env[62814]: INFO nova.scheduler.client.report [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleted allocations for instance 1e259ec6-d31a-453b-87e0-baa446665d56 [ 1082.773227] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba5cf017-09c2-4ae7-8e57-fa99c0041e97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.793933] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1082.793933] env[62814]: value = "task-4294412" [ 1082.793933] env[62814]: _type = "Task" [ 1082.793933] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.807344] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294412, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.922366] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294406, 'name': CreateVM_Task, 'duration_secs': 0.643679} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.922527] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1082.923292] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.923462] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.923779] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1082.924044] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57825e0b-0eb2-4edc-a4d1-4f76a85e1ee0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.928903] env[62814]: DEBUG nova.network.neutron [req-3083047d-4ede-44b2-a5a8-5af77eca4af2 req-161fc7af-9b7f-410a-a79c-458fdf3cfde1 service nova] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updated VIF entry in instance network info cache for port f1881a6c-5678-4b14-9eb9-f3e7f3e41810. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1082.929249] env[62814]: DEBUG nova.network.neutron [req-3083047d-4ede-44b2-a5a8-5af77eca4af2 req-161fc7af-9b7f-410a-a79c-458fdf3cfde1 service nova] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance_info_cache with network_info: [{"id": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "address": "fa:16:3e:35:63:2b", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1881a6c-56", "ovs_interfaceid": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.931464] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1082.931464] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52495f3b-9564-7310-2ea3-0913f4831551" [ 1082.931464] env[62814]: _type = "Task" [ 1082.931464] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.948872] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52495f3b-9564-7310-2ea3-0913f4831551, 'name': SearchDatastore_Task, 'duration_secs': 0.012007} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.949915] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1082.950175] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1082.950404] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.951117] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1082.951331] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1082.952356] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c5be667-3cda-41d1-a26a-e3257a3ba363 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.962859] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1082.962859] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1082.963420] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dcdb1be-ea4c-44ea-82d3-4def9f165154 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.968949] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1082.968949] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f9bff1-a6d7-4e92-0017-f49d6c29f438" [ 1082.968949] env[62814]: _type = "Task" [ 1082.968949] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.976560] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f9bff1-a6d7-4e92-0017-f49d6c29f438, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.075173] env[62814]: DEBUG oslo_vmware.api [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294411, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.355343} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.078790] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1083.079015] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1083.079275] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1083.079383] env[62814]: INFO nova.compute.manager [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1083.079619] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1083.079827] env[62814]: DEBUG oslo_vmware.api [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294410, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.080063] env[62814]: DEBUG nova.compute.manager [-] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1083.080160] env[62814]: DEBUG nova.network.neutron [-] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1083.279502] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6e2642e9-0159-4d0f-b5d6-af7e4b573479 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "1e259ec6-d31a-453b-87e0-baa446665d56" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.903s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1083.306802] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294412, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.433563] env[62814]: DEBUG oslo_concurrency.lockutils [req-3083047d-4ede-44b2-a5a8-5af77eca4af2 req-161fc7af-9b7f-410a-a79c-458fdf3cfde1 service nova] Releasing lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1083.481862] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f9bff1-a6d7-4e92-0017-f49d6c29f438, 'name': SearchDatastore_Task, 'duration_secs': 0.012204} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.482730] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-438d6cbd-f566-48f8-b463-28403ed532f2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.490378] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1083.490378] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d0e6a8-c0b5-237d-0c09-22074c87aa3c" [ 1083.490378] env[62814]: _type = "Task" [ 1083.490378] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.499825] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d0e6a8-c0b5-237d-0c09-22074c87aa3c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.573517] env[62814]: DEBUG oslo_vmware.api [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294410, 'name': PowerOnVM_Task, 'duration_secs': 0.783243} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.573822] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1083.574016] env[62814]: DEBUG nova.compute.manager [None req-672b193e-2666-42d5-a0ce-4d595a366080 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1083.574824] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f67dc4-3098-4e68-a2df-84423dec69db {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.814451] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294412, 'name': ReconfigVM_Task, 'duration_secs': 0.56346} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.814504] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 44ea319c-6ea0-456a-bee6-42133a25d8c1/e7d68632-de75-4206-8f75-4abd879e1d22-rescue.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1083.815504] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d5b237-ea99-48fe-9be9-c3e0a69bfcf7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.862891] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb15f0ed-aa7a-4ec9-a22b-1da814b8e342 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.897163] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1083.897163] env[62814]: value = "task-4294413" [ 1083.897163] env[62814]: _type = "Task" [ 1083.897163] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.913641] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294413, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.003326] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d0e6a8-c0b5-237d-0c09-22074c87aa3c, 'name': SearchDatastore_Task, 'duration_secs': 0.016922} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.006161] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1084.006444] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 7e849b23-226d-4a1d-a85c-cf96964ce034/7e849b23-226d-4a1d-a85c-cf96964ce034.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1084.006980] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b6682a8-30dc-4353-adfb-87d8d7e9ace7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.013988] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1084.013988] env[62814]: value = "task-4294414" [ 1084.013988] env[62814]: _type = "Task" [ 1084.013988] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.022299] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294414, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.233020] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db06437-7a36-4aab-b78b-5703b2a5eadc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.240694] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8482aa2-ed57-48ef-a1a0-418b7538a1a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.277098] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dbfb7e5-bdaf-4405-aef4-be12782d450a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.288741] env[62814]: DEBUG nova.network.neutron [-] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.296821] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3da6d9-b4d4-4747-b15c-c3d3f098e84d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.312356] env[62814]: DEBUG nova.compute.provider_tree [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.408545] env[62814]: DEBUG nova.compute.manager [req-305b4efe-09f4-43e5-8888-51dea6ba44b6 req-e47d04fb-3374-4355-9861-e26234567ff5 service nova] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Received event network-vif-deleted-9fed46ec-04e6-4482-a460-370e4ff04a32 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1084.414958] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294413, 'name': ReconfigVM_Task, 'duration_secs': 0.17166} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.416127] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1084.416127] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e7482ab2-bf56-4df3-bb44-6da17ef9f566 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.423931] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1084.423931] env[62814]: value = "task-4294415" [ 1084.423931] env[62814]: _type = "Task" [ 1084.423931] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.434948] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294415, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.531315] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294414, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.796932] env[62814]: INFO nova.compute.manager [-] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Took 1.72 seconds to deallocate network for instance. [ 1084.799916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.799916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.799916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1084.799916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1084.799916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1084.801768] env[62814]: INFO nova.compute.manager [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Terminating instance [ 1084.816428] env[62814]: DEBUG nova.scheduler.client.report [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1084.938143] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294415, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.026197] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294414, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.685304} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.026480] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 7e849b23-226d-4a1d-a85c-cf96964ce034/7e849b23-226d-4a1d-a85c-cf96964ce034.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1085.026699] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.026967] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b8eef4a6-899e-4176-95c1-0f635d1882af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.038885] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1085.038885] env[62814]: value = "task-4294416" [ 1085.038885] env[62814]: _type = "Task" [ 1085.038885] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.050588] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294416, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.309975] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1085.310631] env[62814]: DEBUG nova.compute.manager [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1085.310826] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1085.311698] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98bcca66-aba8-4910-859d-d04db38b4811 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.319667] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1085.320424] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.601s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1085.320907] env[62814]: DEBUG nova.compute.manager [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1085.323822] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-916e855a-c8e3-4490-83bc-bf66b7c91fc0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.328436] env[62814]: DEBUG oslo_concurrency.lockutils [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.874s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1085.329839] env[62814]: DEBUG nova.objects.instance [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'resources' on Instance uuid 885e0798-2fa6-4f6b-82be-517b6d1168d2 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1085.340902] env[62814]: DEBUG oslo_vmware.api [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1085.340902] env[62814]: value = "task-4294417" [ 1085.340902] env[62814]: _type = "Task" [ 1085.340902] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.354047] env[62814]: DEBUG oslo_vmware.api [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294417, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.438437] env[62814]: DEBUG oslo_vmware.api [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294415, 'name': PowerOnVM_Task, 'duration_secs': 0.519888} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.438735] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1085.442174] env[62814]: DEBUG nova.compute.manager [None req-92b654d9-ff73-4fd9-885e-7371d2690a84 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1085.443011] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd4f55e-488d-4a3e-9854-827027fe5174 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.551041] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294416, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.183979} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.551346] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1085.552763] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8165ef70-e72f-49db-959e-efa37a7524a3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.578592] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 7e849b23-226d-4a1d-a85c-cf96964ce034/7e849b23-226d-4a1d-a85c-cf96964ce034.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.579251] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ddbaaad-7078-4450-bab8-b9226335b55e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.599406] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1085.599406] env[62814]: value = "task-4294418" [ 1085.599406] env[62814]: _type = "Task" [ 1085.599406] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.608343] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294418, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.830639] env[62814]: DEBUG nova.compute.utils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1085.834017] env[62814]: DEBUG nova.compute.manager [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1085.834017] env[62814]: DEBUG nova.network.neutron [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1085.851310] env[62814]: DEBUG oslo_vmware.api [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294417, 'name': PowerOffVM_Task, 'duration_secs': 0.188206} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.851431] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1085.852435] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1085.852435] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9f0e4994-f8ac-4e0c-a2ff-be4a50c441f3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.920746] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1085.920961] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1085.921168] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Deleting the datastore file [datastore2] 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1085.921438] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07801639-6d91-46f6-bdd8-96c6eed2fdb6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.927106] env[62814]: DEBUG nova.policy [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05b537f5305649f19a5ed718b728198c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4622b166b5246639c610559f399b42d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1085.933649] env[62814]: DEBUG oslo_vmware.api [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1085.933649] env[62814]: value = "task-4294420" [ 1085.933649] env[62814]: _type = "Task" [ 1085.933649] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.941742] env[62814]: DEBUG oslo_vmware.api [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294420, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.115314] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294418, 'name': ReconfigVM_Task, 'duration_secs': 0.279593} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.116234] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 7e849b23-226d-4a1d-a85c-cf96964ce034/7e849b23-226d-4a1d-a85c-cf96964ce034.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.117078] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b9f35a9-f3f4-490c-a1ea-99ecf041791d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.125109] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1086.125109] env[62814]: value = "task-4294421" [ 1086.125109] env[62814]: _type = "Task" [ 1086.125109] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.138092] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294421, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.336777] env[62814]: DEBUG nova.compute.manager [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1086.381573] env[62814]: DEBUG nova.network.neutron [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Successfully created port: 5dbe3e06-4f03-4ee9-a3b8-e3906c07663a {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1086.426273] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d31276e-959c-4853-9f9c-b4627fb70fe3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.438712] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70edfa22-2985-4901-8957-4d1019af9767 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.450389] env[62814]: DEBUG oslo_vmware.api [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294420, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186928} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.475532] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1086.475800] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1086.475990] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1086.476182] env[62814]: INFO nova.compute.manager [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Took 1.17 seconds to destroy the instance on the hypervisor. [ 1086.476436] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1086.477997] env[62814]: DEBUG nova.compute.manager [-] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1086.478101] env[62814]: DEBUG nova.network.neutron [-] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1086.480490] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e21994e-e255-4897-8029-1dca124ccf7a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.489060] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be8a039-ed5c-422e-bea8-18ce23488a48 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.509293] env[62814]: DEBUG nova.compute.provider_tree [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1086.558209] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1086.558775] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1086.637967] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294421, 'name': Rename_Task, 'duration_secs': 0.172095} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.638593] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1086.638593] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-11f01218-f52a-4092-80b8-c857a6f39fd9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.645118] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1086.645118] env[62814]: value = "task-4294422" [ 1086.645118] env[62814]: _type = "Task" [ 1086.645118] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.655611] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294422, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.014215] env[62814]: DEBUG nova.scheduler.client.report [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1087.061949] env[62814]: DEBUG nova.compute.manager [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1087.097305] env[62814]: DEBUG nova.compute.manager [req-485fa07d-b5b4-4c3f-bdbd-f46fd157e828 req-17ea6642-5d9f-43c4-a027-fd696bad603c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Received event network-vif-deleted-2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1087.097664] env[62814]: INFO nova.compute.manager [req-485fa07d-b5b4-4c3f-bdbd-f46fd157e828 req-17ea6642-5d9f-43c4-a027-fd696bad603c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Neutron deleted interface 2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f; detaching it from the instance and deleting it from the info cache [ 1087.097707] env[62814]: DEBUG nova.network.neutron [req-485fa07d-b5b4-4c3f-bdbd-f46fd157e828 req-17ea6642-5d9f-43c4-a027-fd696bad603c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.123419] env[62814]: DEBUG nova.compute.manager [req-bc037cbd-c2a1-40d6-9f7e-a433002e8d8b req-5a03a43a-c284-4d07-8a18-161b664f9ef9 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Received event network-changed-d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1087.123721] env[62814]: DEBUG nova.compute.manager [req-bc037cbd-c2a1-40d6-9f7e-a433002e8d8b req-5a03a43a-c284-4d07-8a18-161b664f9ef9 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Refreshing instance network info cache due to event network-changed-d92933dd-ecff-4c85-96cf-69211cb92fa6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1087.123951] env[62814]: DEBUG oslo_concurrency.lockutils [req-bc037cbd-c2a1-40d6-9f7e-a433002e8d8b req-5a03a43a-c284-4d07-8a18-161b664f9ef9 service nova] Acquiring lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.124483] env[62814]: DEBUG oslo_concurrency.lockutils [req-bc037cbd-c2a1-40d6-9f7e-a433002e8d8b req-5a03a43a-c284-4d07-8a18-161b664f9ef9 service nova] Acquired lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1087.125217] env[62814]: DEBUG nova.network.neutron [req-bc037cbd-c2a1-40d6-9f7e-a433002e8d8b req-5a03a43a-c284-4d07-8a18-161b664f9ef9 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Refreshing network info cache for port d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1087.156980] env[62814]: DEBUG oslo_vmware.api [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294422, 'name': PowerOnVM_Task, 'duration_secs': 0.462405} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.158033] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1087.158033] env[62814]: INFO nova.compute.manager [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Took 8.48 seconds to spawn the instance on the hypervisor. [ 1087.158033] env[62814]: DEBUG nova.compute.manager [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1087.159335] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8784f9-73b7-4bdf-827e-f8dfb2a823b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.349236] env[62814]: DEBUG nova.compute.manager [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1087.381357] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1087.381357] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1087.381357] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1087.381357] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1087.381583] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1087.381663] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1087.382873] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1087.382873] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1087.382873] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1087.382873] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1087.382873] env[62814]: DEBUG nova.virt.hardware [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1087.383499] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2e77190-1f56-46bc-b8d4-fffba838db73 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.397044] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1026f2-4573-44d7-b75d-e1863ddb089d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.461620] env[62814]: DEBUG nova.network.neutron [-] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.522749] env[62814]: DEBUG oslo_concurrency.lockutils [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.194s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1087.527107] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.870s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1087.528696] env[62814]: INFO nova.compute.claims [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1087.557099] env[62814]: INFO nova.scheduler.client.report [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleted allocations for instance 885e0798-2fa6-4f6b-82be-517b6d1168d2 [ 1087.594370] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1087.601290] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-efc4f1e9-a1bf-49fa-8c2e-66e72df35a09 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.610934] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0921da87-3b0a-44bd-92e3-598ecde9ba43 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.645336] env[62814]: DEBUG nova.compute.manager [req-485fa07d-b5b4-4c3f-bdbd-f46fd157e828 req-17ea6642-5d9f-43c4-a027-fd696bad603c service nova] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Detach interface failed, port_id=2abec6f6-acb1-4a34-a3ec-4f4ef7f72c1f, reason: Instance 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1087.677645] env[62814]: INFO nova.compute.manager [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Took 30.33 seconds to build instance. [ 1087.912145] env[62814]: DEBUG nova.network.neutron [req-bc037cbd-c2a1-40d6-9f7e-a433002e8d8b req-5a03a43a-c284-4d07-8a18-161b664f9ef9 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updated VIF entry in instance network info cache for port d92933dd-ecff-4c85-96cf-69211cb92fa6. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1087.912145] env[62814]: DEBUG nova.network.neutron [req-bc037cbd-c2a1-40d6-9f7e-a433002e8d8b req-5a03a43a-c284-4d07-8a18-161b664f9ef9 service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updating instance_info_cache with network_info: [{"id": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "address": "fa:16:3e:3e:a0:82", "network": {"id": "6df241c7-9cf0-4d05-baba-469d1bea58ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-62005335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1c7dab0a5e874c3aaa57685a130460fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92933dd-ec", "ovs_interfaceid": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.966762] env[62814]: INFO nova.compute.manager [-] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Took 1.49 seconds to deallocate network for instance. [ 1088.065331] env[62814]: DEBUG oslo_concurrency.lockutils [None req-56a4e9f7-d309-478a-bb06-3fb5086b7145 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "885e0798-2fa6-4f6b-82be-517b6d1168d2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.806s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1088.180747] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7533253b-4140-452d-8d12-bdcf8d5f5752 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.841s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1088.383258] env[62814]: DEBUG nova.network.neutron [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Successfully updated port: 5dbe3e06-4f03-4ee9-a3b8-e3906c07663a {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1088.410545] env[62814]: DEBUG oslo_concurrency.lockutils [req-bc037cbd-c2a1-40d6-9f7e-a433002e8d8b req-5a03a43a-c284-4d07-8a18-161b664f9ef9 service nova] Releasing lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1088.472148] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1088.757671] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquiring lock "e75c8039-9ff8-45f2-8770-894f2307cc14" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1088.758067] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Lock "e75c8039-9ff8-45f2-8770-894f2307cc14" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1088.886987] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquiring lock "refresh_cache-4429ff0c-5353-4346-8bef-2c45d251f1ea" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.886987] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquired lock "refresh_cache-4429ff0c-5353-4346-8bef-2c45d251f1ea" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1088.886987] env[62814]: DEBUG nova.network.neutron [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1088.922559] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5255a36e-af37-42d7-89c4-2b4ad617d221 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.931193] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69755b4a-7981-47ff-b5c9-d37e561268ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.962861] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2223ab73-fbe6-4567-8481-fda94fa46493 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.971977] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b8a12d-c614-4f4b-81ff-1193bedde60a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.989385] env[62814]: DEBUG nova.compute.provider_tree [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1089.259872] env[62814]: DEBUG nova.compute.manager [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1089.390837] env[62814]: DEBUG nova.compute.manager [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Received event network-vif-plugged-5dbe3e06-4f03-4ee9-a3b8-e3906c07663a {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1089.391111] env[62814]: DEBUG oslo_concurrency.lockutils [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] Acquiring lock "4429ff0c-5353-4346-8bef-2c45d251f1ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1089.391466] env[62814]: DEBUG oslo_concurrency.lockutils [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] Lock "4429ff0c-5353-4346-8bef-2c45d251f1ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1089.392940] env[62814]: DEBUG oslo_concurrency.lockutils [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] Lock "4429ff0c-5353-4346-8bef-2c45d251f1ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1089.392940] env[62814]: DEBUG nova.compute.manager [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] No waiting events found dispatching network-vif-plugged-5dbe3e06-4f03-4ee9-a3b8-e3906c07663a {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1089.392940] env[62814]: WARNING nova.compute.manager [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Received unexpected event network-vif-plugged-5dbe3e06-4f03-4ee9-a3b8-e3906c07663a for instance with vm_state building and task_state spawning. [ 1089.392940] env[62814]: DEBUG nova.compute.manager [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Received event network-changed-5dbe3e06-4f03-4ee9-a3b8-e3906c07663a {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1089.392940] env[62814]: DEBUG nova.compute.manager [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Refreshing instance network info cache due to event network-changed-5dbe3e06-4f03-4ee9-a3b8-e3906c07663a. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1089.392940] env[62814]: DEBUG oslo_concurrency.lockutils [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] Acquiring lock "refresh_cache-4429ff0c-5353-4346-8bef-2c45d251f1ea" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.414170] env[62814]: DEBUG nova.compute.manager [req-09de51e2-9abb-402b-b8d1-696b0d014921 req-7c5c116e-6112-4acb-8378-64549c9c2f5e service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Received event network-changed-d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1089.414170] env[62814]: DEBUG nova.compute.manager [req-09de51e2-9abb-402b-b8d1-696b0d014921 req-7c5c116e-6112-4acb-8378-64549c9c2f5e service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Refreshing instance network info cache due to event network-changed-d92933dd-ecff-4c85-96cf-69211cb92fa6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1089.414170] env[62814]: DEBUG oslo_concurrency.lockutils [req-09de51e2-9abb-402b-b8d1-696b0d014921 req-7c5c116e-6112-4acb-8378-64549c9c2f5e service nova] Acquiring lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.414170] env[62814]: DEBUG oslo_concurrency.lockutils [req-09de51e2-9abb-402b-b8d1-696b0d014921 req-7c5c116e-6112-4acb-8378-64549c9c2f5e service nova] Acquired lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1089.414170] env[62814]: DEBUG nova.network.neutron [req-09de51e2-9abb-402b-b8d1-696b0d014921 req-7c5c116e-6112-4acb-8378-64549c9c2f5e service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Refreshing network info cache for port d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1089.432760] env[62814]: DEBUG nova.network.neutron [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1089.497996] env[62814]: DEBUG nova.scheduler.client.report [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1089.732256] env[62814]: DEBUG nova.network.neutron [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Updating instance_info_cache with network_info: [{"id": "5dbe3e06-4f03-4ee9-a3b8-e3906c07663a", "address": "fa:16:3e:37:63:30", "network": {"id": "8be29654-4149-4355-9821-245dc28863bd", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1161337540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4622b166b5246639c610559f399b42d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e49a26b5-7b6b-41fd-8bed-4cd9a6c1a002", "external-id": "nsx-vlan-transportzone-506", "segmentation_id": 506, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5dbe3e06-4f", "ovs_interfaceid": "5dbe3e06-4f03-4ee9-a3b8-e3906c07663a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.788484] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1090.006711] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1090.006804] env[62814]: DEBUG nova.compute.manager [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1090.009273] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.038s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1090.013879] env[62814]: INFO nova.compute.claims [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1090.186282] env[62814]: DEBUG nova.network.neutron [req-09de51e2-9abb-402b-b8d1-696b0d014921 req-7c5c116e-6112-4acb-8378-64549c9c2f5e service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updated VIF entry in instance network info cache for port d92933dd-ecff-4c85-96cf-69211cb92fa6. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1090.186650] env[62814]: DEBUG nova.network.neutron [req-09de51e2-9abb-402b-b8d1-696b0d014921 req-7c5c116e-6112-4acb-8378-64549c9c2f5e service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updating instance_info_cache with network_info: [{"id": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "address": "fa:16:3e:3e:a0:82", "network": {"id": "6df241c7-9cf0-4d05-baba-469d1bea58ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-62005335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1c7dab0a5e874c3aaa57685a130460fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92933dd-ec", "ovs_interfaceid": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.235905] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Releasing lock "refresh_cache-4429ff0c-5353-4346-8bef-2c45d251f1ea" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1090.236105] env[62814]: DEBUG nova.compute.manager [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Instance network_info: |[{"id": "5dbe3e06-4f03-4ee9-a3b8-e3906c07663a", "address": "fa:16:3e:37:63:30", "network": {"id": "8be29654-4149-4355-9821-245dc28863bd", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1161337540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4622b166b5246639c610559f399b42d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e49a26b5-7b6b-41fd-8bed-4cd9a6c1a002", "external-id": "nsx-vlan-transportzone-506", "segmentation_id": 506, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5dbe3e06-4f", "ovs_interfaceid": "5dbe3e06-4f03-4ee9-a3b8-e3906c07663a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1090.236455] env[62814]: DEBUG oslo_concurrency.lockutils [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] Acquired lock "refresh_cache-4429ff0c-5353-4346-8bef-2c45d251f1ea" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1090.236577] env[62814]: DEBUG nova.network.neutron [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Refreshing network info cache for port 5dbe3e06-4f03-4ee9-a3b8-e3906c07663a {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1090.237791] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:63:30', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e49a26b5-7b6b-41fd-8bed-4cd9a6c1a002', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5dbe3e06-4f03-4ee9-a3b8-e3906c07663a', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1090.245552] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Creating folder: Project (b4622b166b5246639c610559f399b42d). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1090.246626] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8e423443-c56f-4c22-9962-6c9682c19850 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.258824] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Created folder: Project (b4622b166b5246639c610559f399b42d) in parent group-v845547. [ 1090.259066] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Creating folder: Instances. Parent ref: group-v845815. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1090.259265] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3ede8955-33a0-4e8f-a14c-b920336d40cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.267816] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Created folder: Instances in parent group-v845815. [ 1090.268089] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1090.268319] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1090.268482] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fe12574-1982-46eb-903b-39ef89b4d1af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.293776] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1090.293776] env[62814]: value = "task-4294425" [ 1090.293776] env[62814]: _type = "Task" [ 1090.293776] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.302675] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294425, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.518960] env[62814]: DEBUG nova.compute.utils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1090.522377] env[62814]: DEBUG nova.compute.manager [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1090.522563] env[62814]: DEBUG nova.network.neutron [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1090.584845] env[62814]: DEBUG nova.policy [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd71116c0cfd4b509e4070ecc84613f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd82993ef7dfa4d2f8f39db3577dd321a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1090.634821] env[62814]: DEBUG nova.compute.manager [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Stashing vm_state: active {{(pid=62814) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1090.690898] env[62814]: DEBUG oslo_concurrency.lockutils [req-09de51e2-9abb-402b-b8d1-696b0d014921 req-7c5c116e-6112-4acb-8378-64549c9c2f5e service nova] Releasing lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1090.804755] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294425, 'name': CreateVM_Task, 'duration_secs': 0.49291} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.805095] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1090.805599] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1090.805766] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1090.806641] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1090.807688] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a891323-551c-4e8b-9d9d-69f0d2f2220f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.815078] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for the task: (returnval){ [ 1090.815078] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52690a4b-a42a-8ccb-f2e3-d4ebb49a8fc3" [ 1090.815078] env[62814]: _type = "Task" [ 1090.815078] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.829115] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52690a4b-a42a-8ccb-f2e3-d4ebb49a8fc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.026610] env[62814]: DEBUG nova.compute.manager [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1091.148531] env[62814]: DEBUG nova.network.neutron [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Updated VIF entry in instance network info cache for port 5dbe3e06-4f03-4ee9-a3b8-e3906c07663a. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1091.149392] env[62814]: DEBUG nova.network.neutron [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Updating instance_info_cache with network_info: [{"id": "5dbe3e06-4f03-4ee9-a3b8-e3906c07663a", "address": "fa:16:3e:37:63:30", "network": {"id": "8be29654-4149-4355-9821-245dc28863bd", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1161337540-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4622b166b5246639c610559f399b42d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e49a26b5-7b6b-41fd-8bed-4cd9a6c1a002", "external-id": "nsx-vlan-transportzone-506", "segmentation_id": 506, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5dbe3e06-4f", "ovs_interfaceid": "5dbe3e06-4f03-4ee9-a3b8-e3906c07663a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.163140] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1091.193994] env[62814]: DEBUG nova.network.neutron [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Successfully created port: bb09100c-d462-4d19-9fd5-1726001e8e55 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1091.332747] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52690a4b-a42a-8ccb-f2e3-d4ebb49a8fc3, 'name': SearchDatastore_Task, 'duration_secs': 0.021397} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.333109] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1091.333542] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1091.334213] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.334213] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1091.334213] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1091.334464] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cfedb6b6-e702-4c39-8c41-5a383d4ff07b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.359161] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1091.359456] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1091.360192] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ea4074c-f402-443d-8d1a-b1a8bdaa6596 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.367122] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for the task: (returnval){ [ 1091.367122] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fd7935-c276-9266-7cce-213325a2fa4e" [ 1091.367122] env[62814]: _type = "Task" [ 1091.367122] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.378124] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fd7935-c276-9266-7cce-213325a2fa4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.483839] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1a6242-b900-4b97-a713-e39bf022898c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.488876] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4420374-1225-4e86-b6fd-23ff45b2b4a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.522258] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5929a8de-2e6d-4525-9f8e-b2b643b80f27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.530316] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35807561-e000-4493-af48-5513fa82eec5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.548718] env[62814]: DEBUG nova.compute.provider_tree [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.654140] env[62814]: DEBUG oslo_concurrency.lockutils [req-244e3cb5-6494-400d-8388-b48813176ef9 req-8d7a9928-60a8-43c0-8470-10831a88f169 service nova] Releasing lock "refresh_cache-4429ff0c-5353-4346-8bef-2c45d251f1ea" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1091.777495] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1091.777731] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1091.877833] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52fd7935-c276-9266-7cce-213325a2fa4e, 'name': SearchDatastore_Task, 'duration_secs': 0.029214} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.878674] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57693827-4b9c-4cad-8690-d3dbdc588a6c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.884203] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for the task: (returnval){ [ 1091.884203] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522a2bb6-2438-2162-e22c-ba22112681e4" [ 1091.884203] env[62814]: _type = "Task" [ 1091.884203] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.892601] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522a2bb6-2438-2162-e22c-ba22112681e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.976414] env[62814]: DEBUG nova.compute.manager [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Received event network-changed-d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1091.976414] env[62814]: DEBUG nova.compute.manager [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Refreshing instance network info cache due to event network-changed-d92933dd-ecff-4c85-96cf-69211cb92fa6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1091.976414] env[62814]: DEBUG oslo_concurrency.lockutils [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] Acquiring lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.976414] env[62814]: DEBUG oslo_concurrency.lockutils [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] Acquired lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1091.976414] env[62814]: DEBUG nova.network.neutron [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Refreshing network info cache for port d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1092.043940] env[62814]: DEBUG nova.compute.manager [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1092.051588] env[62814]: DEBUG nova.scheduler.client.report [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1092.073710] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1092.073978] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1092.074157] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1092.074797] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1092.074797] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1092.074797] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1092.074797] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1092.074995] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1092.075139] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1092.075326] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1092.075535] env[62814]: DEBUG nova.virt.hardware [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1092.076694] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70dd7557-a23a-42e1-9798-d1205dd97ad6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.084811] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-590bc2e1-3cd5-4611-9f22-2d111074f4ed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.280541] env[62814]: DEBUG nova.compute.manager [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1092.396673] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522a2bb6-2438-2162-e22c-ba22112681e4, 'name': SearchDatastore_Task, 'duration_secs': 0.012446} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.396673] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1092.396885] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 4429ff0c-5353-4346-8bef-2c45d251f1ea/4429ff0c-5353-4346-8bef-2c45d251f1ea.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1092.397245] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2253284d-afbf-48d2-ab50-40b00970b311 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.406275] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for the task: (returnval){ [ 1092.406275] env[62814]: value = "task-4294426" [ 1092.406275] env[62814]: _type = "Task" [ 1092.406275] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.414964] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294426, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.557762] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.557877] env[62814]: DEBUG nova.compute.manager [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1092.560887] env[62814]: DEBUG oslo_concurrency.lockutils [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 23.169s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.689968] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquiring lock "44ea319c-6ea0-456a-bee6-42133a25d8c1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.690334] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lock "44ea319c-6ea0-456a-bee6-42133a25d8c1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.690558] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquiring lock "44ea319c-6ea0-456a-bee6-42133a25d8c1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.690776] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lock "44ea319c-6ea0-456a-bee6-42133a25d8c1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1092.690983] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lock "44ea319c-6ea0-456a-bee6-42133a25d8c1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1092.693495] env[62814]: INFO nova.compute.manager [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Terminating instance [ 1092.810006] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1092.890929] env[62814]: DEBUG nova.network.neutron [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updated VIF entry in instance network info cache for port d92933dd-ecff-4c85-96cf-69211cb92fa6. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1092.891350] env[62814]: DEBUG nova.network.neutron [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updating instance_info_cache with network_info: [{"id": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "address": "fa:16:3e:3e:a0:82", "network": {"id": "6df241c7-9cf0-4d05-baba-469d1bea58ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-62005335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1c7dab0a5e874c3aaa57685a130460fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92933dd-ec", "ovs_interfaceid": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.932128] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294426, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50608} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.932128] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 4429ff0c-5353-4346-8bef-2c45d251f1ea/4429ff0c-5353-4346-8bef-2c45d251f1ea.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1092.932128] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1092.932128] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-99ff628b-ce02-4183-a793-20efa8adfeb8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.945276] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for the task: (returnval){ [ 1092.945276] env[62814]: value = "task-4294427" [ 1092.945276] env[62814]: _type = "Task" [ 1092.945276] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.958021] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294427, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.007504] env[62814]: DEBUG nova.compute.manager [req-d1788ec3-953d-4586-afd0-1cfa737a108e req-4244413c-8337-4fe7-9c6f-ffc975eb23b9 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Received event network-vif-plugged-bb09100c-d462-4d19-9fd5-1726001e8e55 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1093.007504] env[62814]: DEBUG oslo_concurrency.lockutils [req-d1788ec3-953d-4586-afd0-1cfa737a108e req-4244413c-8337-4fe7-9c6f-ffc975eb23b9 service nova] Acquiring lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1093.007504] env[62814]: DEBUG oslo_concurrency.lockutils [req-d1788ec3-953d-4586-afd0-1cfa737a108e req-4244413c-8337-4fe7-9c6f-ffc975eb23b9 service nova] Lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1093.007504] env[62814]: DEBUG oslo_concurrency.lockutils [req-d1788ec3-953d-4586-afd0-1cfa737a108e req-4244413c-8337-4fe7-9c6f-ffc975eb23b9 service nova] Lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1093.007504] env[62814]: DEBUG nova.compute.manager [req-d1788ec3-953d-4586-afd0-1cfa737a108e req-4244413c-8337-4fe7-9c6f-ffc975eb23b9 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] No waiting events found dispatching network-vif-plugged-bb09100c-d462-4d19-9fd5-1726001e8e55 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1093.007504] env[62814]: WARNING nova.compute.manager [req-d1788ec3-953d-4586-afd0-1cfa737a108e req-4244413c-8337-4fe7-9c6f-ffc975eb23b9 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Received unexpected event network-vif-plugged-bb09100c-d462-4d19-9fd5-1726001e8e55 for instance with vm_state building and task_state spawning. [ 1093.064386] env[62814]: DEBUG nova.compute.utils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1093.069309] env[62814]: DEBUG nova.compute.manager [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1093.069490] env[62814]: DEBUG nova.network.neutron [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1093.145947] env[62814]: DEBUG nova.policy [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '914cba82d20a43c28f31e33e09f61459', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '67cc9d091e094518a5ba7b78708698fe', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1093.161079] env[62814]: DEBUG nova.network.neutron [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Successfully updated port: bb09100c-d462-4d19-9fd5-1726001e8e55 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1093.201695] env[62814]: DEBUG nova.compute.manager [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1093.201918] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1093.202995] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf775319-494a-4d60-80b7-f86677f96eec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.212897] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1093.213239] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5940be7-a88d-431b-8481-cf419ede2144 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.225265] env[62814]: DEBUG oslo_vmware.api [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1093.225265] env[62814]: value = "task-4294428" [ 1093.225265] env[62814]: _type = "Task" [ 1093.225265] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.232923] env[62814]: DEBUG oslo_vmware.api [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294428, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.395741] env[62814]: DEBUG oslo_concurrency.lockutils [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] Releasing lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1093.395999] env[62814]: DEBUG nova.compute.manager [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Received event network-changed-d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1093.396472] env[62814]: DEBUG nova.compute.manager [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Refreshing instance network info cache due to event network-changed-d92933dd-ecff-4c85-96cf-69211cb92fa6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1093.396646] env[62814]: DEBUG oslo_concurrency.lockutils [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] Acquiring lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.396876] env[62814]: DEBUG oslo_concurrency.lockutils [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] Acquired lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.396961] env[62814]: DEBUG nova.network.neutron [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Refreshing network info cache for port d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1093.410826] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa84e44a-b09e-48d1-a86b-c4c55f745af0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.419574] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b163236a-752e-45fe-8040-4377dba220f3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.455785] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da63e9f-c221-4622-b6bf-e6a44f3ff609 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.458792] env[62814]: DEBUG nova.network.neutron [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Successfully created port: 8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1093.467723] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294427, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.154143} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.468056] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1093.469247] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d707b7f-ae95-492e-96cd-864e38b88c58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.473298] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f56d5ca1-b3f4-4557-b19d-2df1f107faf3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.486041] env[62814]: DEBUG nova.compute.provider_tree [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.508488] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 4429ff0c-5353-4346-8bef-2c45d251f1ea/4429ff0c-5353-4346-8bef-2c45d251f1ea.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1093.509846] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51a728cc-11c4-4d62-96cd-e4c474cd472d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.530370] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for the task: (returnval){ [ 1093.530370] env[62814]: value = "task-4294429" [ 1093.530370] env[62814]: _type = "Task" [ 1093.530370] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.540321] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294429, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.571659] env[62814]: DEBUG nova.compute.manager [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1093.661080] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "refresh_cache-201cf7be-f81f-4eba-ae65-a370ad8e48c3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.661251] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "refresh_cache-201cf7be-f81f-4eba-ae65-a370ad8e48c3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1093.661409] env[62814]: DEBUG nova.network.neutron [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1093.737230] env[62814]: DEBUG oslo_vmware.api [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294428, 'name': PowerOffVM_Task, 'duration_secs': 0.392778} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.737230] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1093.737230] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1093.737550] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1fcc8026-4f25-4f22-8bef-8a12d0c4fc22 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.978348] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1093.978621] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1093.978670] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Deleting the datastore file [datastore2] 44ea319c-6ea0-456a-bee6-42133a25d8c1 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1093.978935] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee9fc08a-ca77-4022-b1a9-343dedc720fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.985990] env[62814]: DEBUG oslo_vmware.api [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for the task: (returnval){ [ 1093.985990] env[62814]: value = "task-4294431" [ 1093.985990] env[62814]: _type = "Task" [ 1093.985990] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.995249] env[62814]: DEBUG oslo_vmware.api [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294431, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.010705] env[62814]: DEBUG nova.scheduler.client.report [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1094.043310] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294429, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.168325] env[62814]: DEBUG nova.network.neutron [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updated VIF entry in instance network info cache for port d92933dd-ecff-4c85-96cf-69211cb92fa6. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1094.169037] env[62814]: DEBUG nova.network.neutron [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updating instance_info_cache with network_info: [{"id": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "address": "fa:16:3e:3e:a0:82", "network": {"id": "6df241c7-9cf0-4d05-baba-469d1bea58ee", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-62005335-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "1c7dab0a5e874c3aaa57685a130460fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd92933dd-ec", "ovs_interfaceid": "d92933dd-ecff-4c85-96cf-69211cb92fa6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.194576] env[62814]: DEBUG nova.network.neutron [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1094.332927] env[62814]: DEBUG nova.network.neutron [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Updating instance_info_cache with network_info: [{"id": "bb09100c-d462-4d19-9fd5-1726001e8e55", "address": "fa:16:3e:1d:27:0c", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb09100c-d4", "ovs_interfaceid": "bb09100c-d462-4d19-9fd5-1726001e8e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1094.496092] env[62814]: DEBUG oslo_vmware.api [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Task: {'id': task-4294431, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.49766} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.496228] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1094.496404] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1094.496587] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1094.496757] env[62814]: INFO nova.compute.manager [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Took 1.29 seconds to destroy the instance on the hypervisor. [ 1094.496997] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1094.497213] env[62814]: DEBUG nova.compute.manager [-] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1094.497302] env[62814]: DEBUG nova.network.neutron [-] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1094.541783] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294429, 'name': ReconfigVM_Task, 'duration_secs': 0.64189} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.542178] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 4429ff0c-5353-4346-8bef-2c45d251f1ea/4429ff0c-5353-4346-8bef-2c45d251f1ea.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1094.542839] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dd1db12a-33ec-4a49-b083-21c111183391 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.549863] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for the task: (returnval){ [ 1094.549863] env[62814]: value = "task-4294432" [ 1094.549863] env[62814]: _type = "Task" [ 1094.549863] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.557472] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294432, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.580707] env[62814]: DEBUG nova.compute.manager [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1094.610237] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1094.611781] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1094.611781] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1094.611781] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1094.611781] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1094.611781] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1094.611781] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1094.611781] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1094.611781] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1094.612103] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1094.612103] env[62814]: DEBUG nova.virt.hardware [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1094.613428] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c65e526-7d44-4b98-baf4-0af3903ffb8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.621175] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b81c727c-3d49-46a2-aaf4-da36dba06a42 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.675344] env[62814]: DEBUG oslo_concurrency.lockutils [req-33b1cc31-dbbc-456f-8e4a-08f1449318f4 req-ccaae9fd-7a20-4a02-9af6-10b0ec400b1a service nova] Releasing lock "refresh_cache-44ea319c-6ea0-456a-bee6-42133a25d8c1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1094.837475] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "refresh_cache-201cf7be-f81f-4eba-ae65-a370ad8e48c3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1094.837785] env[62814]: DEBUG nova.compute.manager [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Instance network_info: |[{"id": "bb09100c-d462-4d19-9fd5-1726001e8e55", "address": "fa:16:3e:1d:27:0c", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb09100c-d4", "ovs_interfaceid": "bb09100c-d462-4d19-9fd5-1726001e8e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1094.838843] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:27:0c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bb09100c-d462-4d19-9fd5-1726001e8e55', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1094.845977] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1094.846255] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1094.847221] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e759ab96-ae6c-45e2-b1bf-191dea5f7061 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.867652] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1094.867652] env[62814]: value = "task-4294433" [ 1094.867652] env[62814]: _type = "Task" [ 1094.867652] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.878334] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294433, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.956115] env[62814]: DEBUG nova.compute.manager [req-4db6a225-ecb5-4aae-bfb6-7647f4c55f85 req-bb3c7ce7-32d8-44b6-9e40-b4ca828b0ffc service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Received event network-vif-deleted-d92933dd-ecff-4c85-96cf-69211cb92fa6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1094.956335] env[62814]: INFO nova.compute.manager [req-4db6a225-ecb5-4aae-bfb6-7647f4c55f85 req-bb3c7ce7-32d8-44b6-9e40-b4ca828b0ffc service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Neutron deleted interface d92933dd-ecff-4c85-96cf-69211cb92fa6; detaching it from the instance and deleting it from the info cache [ 1094.956530] env[62814]: DEBUG nova.network.neutron [req-4db6a225-ecb5-4aae-bfb6-7647f4c55f85 req-bb3c7ce7-32d8-44b6-9e40-b4ca828b0ffc service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.024055] env[62814]: DEBUG oslo_concurrency.lockutils [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.463s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1095.028778] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.219s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1095.029082] env[62814]: DEBUG nova.objects.instance [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lazy-loading 'resources' on Instance uuid cd037f6e-fddd-4389-b6d0-144b798537bf {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.045973] env[62814]: DEBUG nova.compute.manager [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Received event network-changed-bb09100c-d462-4d19-9fd5-1726001e8e55 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1095.046197] env[62814]: DEBUG nova.compute.manager [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Refreshing instance network info cache due to event network-changed-bb09100c-d462-4d19-9fd5-1726001e8e55. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1095.047700] env[62814]: DEBUG oslo_concurrency.lockutils [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] Acquiring lock "refresh_cache-201cf7be-f81f-4eba-ae65-a370ad8e48c3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.047884] env[62814]: DEBUG oslo_concurrency.lockutils [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] Acquired lock "refresh_cache-201cf7be-f81f-4eba-ae65-a370ad8e48c3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.048095] env[62814]: DEBUG nova.network.neutron [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Refreshing network info cache for port bb09100c-d462-4d19-9fd5-1726001e8e55 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1095.060854] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294432, 'name': Rename_Task, 'duration_secs': 0.22996} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.061699] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1095.062470] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-222ea681-bdf9-4eda-8a76-bac995ea0149 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.069953] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for the task: (returnval){ [ 1095.069953] env[62814]: value = "task-4294434" [ 1095.069953] env[62814]: _type = "Task" [ 1095.069953] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.079855] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294434, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.108164] env[62814]: DEBUG nova.network.neutron [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Successfully updated port: 8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1095.337545] env[62814]: DEBUG nova.network.neutron [-] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.377551] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294433, 'name': CreateVM_Task, 'duration_secs': 0.383337} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.377734] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1095.378509] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.378619] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.378944] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1095.379222] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c877204-a0f3-49e7-9f18-842060e00c9b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.383923] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1095.383923] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526e247f-a3ab-f53b-234c-52b39153f2e9" [ 1095.383923] env[62814]: _type = "Task" [ 1095.383923] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.391559] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526e247f-a3ab-f53b-234c-52b39153f2e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.461330] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c40bc2d0-c28f-475a-aa87-af83dd4bccb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.470255] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb0f0f6f-9aec-4c96-bf39-9fd96904291b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.501084] env[62814]: DEBUG nova.compute.manager [req-4db6a225-ecb5-4aae-bfb6-7647f4c55f85 req-bb3c7ce7-32d8-44b6-9e40-b4ca828b0ffc service nova] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Detach interface failed, port_id=d92933dd-ecff-4c85-96cf-69211cb92fa6, reason: Instance 44ea319c-6ea0-456a-bee6-42133a25d8c1 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1095.582645] env[62814]: DEBUG oslo_vmware.api [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294434, 'name': PowerOnVM_Task, 'duration_secs': 0.483924} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.585010] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1095.585258] env[62814]: INFO nova.compute.manager [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Took 8.24 seconds to spawn the instance on the hypervisor. [ 1095.585441] env[62814]: DEBUG nova.compute.manager [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1095.586414] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c61ea6-7067-4bab-a5f7-39ad3786ff6b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.591305] env[62814]: INFO nova.scheduler.client.report [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleted allocation for migration 953ceb03-e52d-4c0a-9b6d-70bce2963e8d [ 1095.610608] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.610857] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquired lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.611122] env[62814]: DEBUG nova.network.neutron [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1095.793909] env[62814]: DEBUG nova.network.neutron [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Updated VIF entry in instance network info cache for port bb09100c-d462-4d19-9fd5-1726001e8e55. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1095.793994] env[62814]: DEBUG nova.network.neutron [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Updating instance_info_cache with network_info: [{"id": "bb09100c-d462-4d19-9fd5-1726001e8e55", "address": "fa:16:3e:1d:27:0c", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbb09100c-d4", "ovs_interfaceid": "bb09100c-d462-4d19-9fd5-1726001e8e55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.840474] env[62814]: INFO nova.compute.manager [-] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Took 1.34 seconds to deallocate network for instance. [ 1095.855196] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c579e80-e24e-4970-a247-881c6fe78742 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.866121] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8430e2e-2000-43aa-8df2-adca2ee651f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.900282] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-993ee9e0-7041-49e7-bf35-27f985b14e6e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.908783] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526e247f-a3ab-f53b-234c-52b39153f2e9, 'name': SearchDatastore_Task, 'duration_secs': 0.009627} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.910927] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1095.911164] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1095.911396] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.911546] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1095.911725] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1095.912053] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99adcdb3-79e7-43e1-a2cd-369b3e95ad13 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.914890] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e1764f5-6a47-4ff5-861b-2785f8c0847b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.931981] env[62814]: DEBUG nova.compute.provider_tree [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.936531] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1095.936531] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1095.936531] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8e02676-9e6e-4cfe-b07a-83362a511c9f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.941210] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1095.941210] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a488e2-386b-837c-e705-61a9530684d9" [ 1095.941210] env[62814]: _type = "Task" [ 1095.941210] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.951379] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a488e2-386b-837c-e705-61a9530684d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.102050] env[62814]: DEBUG oslo_concurrency.lockutils [None req-070cb73e-0fca-4218-b008-600c2abbf0d0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 30.324s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.110064] env[62814]: INFO nova.compute.manager [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Took 32.30 seconds to build instance. [ 1096.161892] env[62814]: DEBUG nova.network.neutron [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1096.292838] env[62814]: DEBUG nova.network.neutron [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Updating instance_info_cache with network_info: [{"id": "8fb80dc8-60f2-403a-99f7-4c77d3431369", "address": "fa:16:3e:4d:02:44", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fb80dc8-60", "ovs_interfaceid": "8fb80dc8-60f2-403a-99f7-4c77d3431369", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1096.296576] env[62814]: DEBUG oslo_concurrency.lockutils [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] Releasing lock "refresh_cache-201cf7be-f81f-4eba-ae65-a370ad8e48c3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.296897] env[62814]: DEBUG nova.compute.manager [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Received event network-vif-plugged-8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1096.297158] env[62814]: DEBUG oslo_concurrency.lockutils [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] Acquiring lock "08c20588-2988-4976-8631-53cf447b4877-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.297980] env[62814]: DEBUG oslo_concurrency.lockutils [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] Lock "08c20588-2988-4976-8631-53cf447b4877-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.298269] env[62814]: DEBUG oslo_concurrency.lockutils [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] Lock "08c20588-2988-4976-8631-53cf447b4877-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.298479] env[62814]: DEBUG nova.compute.manager [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] No waiting events found dispatching network-vif-plugged-8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1096.298665] env[62814]: WARNING nova.compute.manager [req-727ccee0-5e8f-4a40-8a31-0be8260c8589 req-7111d2ed-cc7f-45a0-8389-fb5b234dab09 service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Received unexpected event network-vif-plugged-8fb80dc8-60f2-403a-99f7-4c77d3431369 for instance with vm_state building and task_state spawning. [ 1096.351750] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.437073] env[62814]: DEBUG nova.scheduler.client.report [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1096.451663] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a488e2-386b-837c-e705-61a9530684d9, 'name': SearchDatastore_Task, 'duration_secs': 0.009031} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.452534] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93d9071b-705e-4c1d-a449-988639fac388 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.458422] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1096.458422] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d79a98-400e-060a-f3db-f15928668383" [ 1096.458422] env[62814]: _type = "Task" [ 1096.458422] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.467074] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d79a98-400e-060a-f3db-f15928668383, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.612881] env[62814]: DEBUG oslo_concurrency.lockutils [None req-40323c1b-f2e3-49b5-abb0-c18191bfdaef tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lock "4429ff0c-5353-4346-8bef-2c45d251f1ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.814s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.767809] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "59d21ef7-df97-49ac-9329-4c18df6dd087" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.768149] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.768409] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "59d21ef7-df97-49ac-9329-4c18df6dd087-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1096.768652] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.768831] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.771035] env[62814]: INFO nova.compute.manager [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Terminating instance [ 1096.795810] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Releasing lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.796125] env[62814]: DEBUG nova.compute.manager [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Instance network_info: |[{"id": "8fb80dc8-60f2-403a-99f7-4c77d3431369", "address": "fa:16:3e:4d:02:44", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fb80dc8-60", "ovs_interfaceid": "8fb80dc8-60f2-403a-99f7-4c77d3431369", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1096.796513] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:02:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c883fb98-d172-4510-8cf4-07aafdf771af', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8fb80dc8-60f2-403a-99f7-4c77d3431369', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1096.804250] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1096.804704] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08c20588-2988-4976-8631-53cf447b4877] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1096.804934] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c77d6e1-44ef-4fc2-8cbd-6aef5a2cffbe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.824515] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1096.824515] env[62814]: value = "task-4294435" [ 1096.824515] env[62814]: _type = "Task" [ 1096.824515] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.834432] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294435, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.942259] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.913s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1096.945103] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.465s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1096.945103] env[62814]: DEBUG nova.objects.instance [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lazy-loading 'resources' on Instance uuid b4ee4cee-3298-4955-8375-8ca8c04b2f9f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1096.969386] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d79a98-400e-060a-f3db-f15928668383, 'name': SearchDatastore_Task, 'duration_secs': 0.009754} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.969767] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1096.969957] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 201cf7be-f81f-4eba-ae65-a370ad8e48c3/201cf7be-f81f-4eba-ae65-a370ad8e48c3.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1096.970230] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1df5fb8a-c9f6-42e6-80df-45446df56497 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.975267] env[62814]: INFO nova.scheduler.client.report [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleted allocations for instance cd037f6e-fddd-4389-b6d0-144b798537bf [ 1096.978037] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1096.978037] env[62814]: value = "task-4294436" [ 1096.978037] env[62814]: _type = "Task" [ 1096.978037] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.988379] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294436, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.100360] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquiring lock "4429ff0c-5353-4346-8bef-2c45d251f1ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1097.102714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lock "4429ff0c-5353-4346-8bef-2c45d251f1ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1097.102714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquiring lock "4429ff0c-5353-4346-8bef-2c45d251f1ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1097.102714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lock "4429ff0c-5353-4346-8bef-2c45d251f1ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1097.102714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lock "4429ff0c-5353-4346-8bef-2c45d251f1ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.103733] env[62814]: INFO nova.compute.manager [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Terminating instance [ 1097.113098] env[62814]: DEBUG nova.compute.manager [req-19497080-fc3a-44de-bc3a-fd9002c84438 req-94a22668-d6d3-4895-bc47-4b5cb543ef2c service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Received event network-changed-8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1097.113098] env[62814]: DEBUG nova.compute.manager [req-19497080-fc3a-44de-bc3a-fd9002c84438 req-94a22668-d6d3-4895-bc47-4b5cb543ef2c service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Refreshing instance network info cache due to event network-changed-8fb80dc8-60f2-403a-99f7-4c77d3431369. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1097.113488] env[62814]: DEBUG oslo_concurrency.lockutils [req-19497080-fc3a-44de-bc3a-fd9002c84438 req-94a22668-d6d3-4895-bc47-4b5cb543ef2c service nova] Acquiring lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.113533] env[62814]: DEBUG oslo_concurrency.lockutils [req-19497080-fc3a-44de-bc3a-fd9002c84438 req-94a22668-d6d3-4895-bc47-4b5cb543ef2c service nova] Acquired lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.113702] env[62814]: DEBUG nova.network.neutron [req-19497080-fc3a-44de-bc3a-fd9002c84438 req-94a22668-d6d3-4895-bc47-4b5cb543ef2c service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Refreshing network info cache for port 8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1097.274745] env[62814]: DEBUG nova.compute.manager [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1097.274999] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1097.276085] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a3ee4e1-0d34-458c-b7ca-86cfcb7a3e8f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.284608] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.284940] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-851808ed-bc59-4449-b3d0-10b7977689be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.292207] env[62814]: DEBUG oslo_vmware.api [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1097.292207] env[62814]: value = "task-4294437" [ 1097.292207] env[62814]: _type = "Task" [ 1097.292207] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.301421] env[62814]: DEBUG oslo_vmware.api [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294437, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.335394] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294435, 'name': CreateVM_Task, 'duration_secs': 0.333049} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.335687] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08c20588-2988-4976-8631-53cf447b4877] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1097.336430] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.336667] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.337251] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1097.337320] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2074ce02-8d05-4cd5-99df-e30cf08e7b95 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.342088] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1097.342088] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52402a6e-2948-6365-2dc7-136a7ae1f5b0" [ 1097.342088] env[62814]: _type = "Task" [ 1097.342088] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.350417] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52402a6e-2948-6365-2dc7-136a7ae1f5b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.496425] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294436, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.496875] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f1de9410-49ea-4589-8990-1162590cb5e6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "cd037f6e-fddd-4389-b6d0-144b798537bf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.242s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1097.608956] env[62814]: DEBUG nova.compute.manager [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1097.608956] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1097.609404] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf0a3d4-43cd-4ac9-9c6d-a618e00d2d93 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.621297] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1097.621552] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e85f5a1-a400-4057-82b1-f478be7070b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.631246] env[62814]: DEBUG oslo_vmware.api [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for the task: (returnval){ [ 1097.631246] env[62814]: value = "task-4294438" [ 1097.631246] env[62814]: _type = "Task" [ 1097.631246] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.641398] env[62814]: DEBUG oslo_vmware.api [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294438, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.798559] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f83c74a-6a3d-47c1-a6e6-ee0caf5ebc5a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.805977] env[62814]: DEBUG oslo_vmware.api [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294437, 'name': PowerOffVM_Task, 'duration_secs': 0.290739} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.807921] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1097.808117] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1097.808413] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9f885ef-8367-4a19-ad66-08ea0c01075b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.813694] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd36a3c-a5b2-43be-b765-c88256e0b593 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.852417] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31c2a3a-e7d0-426c-b2ac-8362f9fb5f6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.864460] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52402a6e-2948-6365-2dc7-136a7ae1f5b0, 'name': SearchDatastore_Task, 'duration_secs': 0.057484} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.866727] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1097.867031] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1097.867295] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.867450] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1097.867675] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1097.868085] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8218a263-1818-4bfa-800d-9d16b024fcd1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.871453] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088ecc53-9810-4e24-a5ac-74e8d407ab91 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.878167] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1097.878364] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1097.878573] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleting the datastore file [datastore2] 59d21ef7-df97-49ac-9329-4c18df6dd087 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1097.878842] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2c897c9f-dfed-47e2-9a59-ea560fecc650 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.888450] env[62814]: DEBUG nova.compute.provider_tree [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.890369] env[62814]: DEBUG nova.network.neutron [req-19497080-fc3a-44de-bc3a-fd9002c84438 req-94a22668-d6d3-4895-bc47-4b5cb543ef2c service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Updated VIF entry in instance network info cache for port 8fb80dc8-60f2-403a-99f7-4c77d3431369. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1097.890685] env[62814]: DEBUG nova.network.neutron [req-19497080-fc3a-44de-bc3a-fd9002c84438 req-94a22668-d6d3-4895-bc47-4b5cb543ef2c service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Updating instance_info_cache with network_info: [{"id": "8fb80dc8-60f2-403a-99f7-4c77d3431369", "address": "fa:16:3e:4d:02:44", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fb80dc8-60", "ovs_interfaceid": "8fb80dc8-60f2-403a-99f7-4c77d3431369", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.895124] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1097.895124] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1097.895124] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9f2cd71-001a-4460-a89f-e9a2bed41625 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.897222] env[62814]: DEBUG oslo_vmware.api [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1097.897222] env[62814]: value = "task-4294440" [ 1097.897222] env[62814]: _type = "Task" [ 1097.897222] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.902344] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1097.902344] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5211e3fd-4ede-a6a7-b412-034b1b789c0c" [ 1097.902344] env[62814]: _type = "Task" [ 1097.902344] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.908623] env[62814]: DEBUG oslo_vmware.api [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294440, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.914099] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5211e3fd-4ede-a6a7-b412-034b1b789c0c, 'name': SearchDatastore_Task, 'duration_secs': 0.008907} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.914431] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b089620-a773-447a-8125-acfbbbd126b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.922219] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1097.922219] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52be264d-40c4-e9d5-c47e-13c80e557337" [ 1097.922219] env[62814]: _type = "Task" [ 1097.922219] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.929199] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52be264d-40c4-e9d5-c47e-13c80e557337, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.990545] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294436, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568687} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.990861] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 201cf7be-f81f-4eba-ae65-a370ad8e48c3/201cf7be-f81f-4eba-ae65-a370ad8e48c3.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1097.991134] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1097.991436] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af94de9f-21ca-42ea-84d6-1c23bd10e1b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.998571] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1097.998571] env[62814]: value = "task-4294441" [ 1097.998571] env[62814]: _type = "Task" [ 1097.998571] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.007593] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294441, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.140839] env[62814]: DEBUG oslo_vmware.api [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294438, 'name': PowerOffVM_Task, 'duration_secs': 0.166272} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.141135] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1098.141184] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1098.142047] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-456fb98a-d93b-424e-9ec2-11bcf31d4784 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.198989] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1098.198989] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1098.199134] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Deleting the datastore file [datastore2] 4429ff0c-5353-4346-8bef-2c45d251f1ea {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.199449] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-791fa3cc-6495-4d33-a739-80eeb80a5b91 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.206069] env[62814]: DEBUG oslo_vmware.api [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for the task: (returnval){ [ 1098.206069] env[62814]: value = "task-4294443" [ 1098.206069] env[62814]: _type = "Task" [ 1098.206069] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.214458] env[62814]: DEBUG oslo_vmware.api [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294443, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.394468] env[62814]: DEBUG oslo_concurrency.lockutils [req-19497080-fc3a-44de-bc3a-fd9002c84438 req-94a22668-d6d3-4895-bc47-4b5cb543ef2c service nova] Releasing lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.395472] env[62814]: DEBUG nova.scheduler.client.report [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1098.408043] env[62814]: DEBUG oslo_vmware.api [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294440, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.205643} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.408295] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.408477] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1098.408656] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1098.408985] env[62814]: INFO nova.compute.manager [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1098.409075] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1098.410045] env[62814]: DEBUG nova.compute.manager [-] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1098.410045] env[62814]: DEBUG nova.network.neutron [-] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.431223] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52be264d-40c4-e9d5-c47e-13c80e557337, 'name': SearchDatastore_Task, 'duration_secs': 0.010606} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.431223] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1098.431223] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 08c20588-2988-4976-8631-53cf447b4877/08c20588-2988-4976-8631-53cf447b4877.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1098.431223] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d3fc17ee-1bc5-49e1-b10c-f3873d6c3e6e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.436842] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1098.436842] env[62814]: value = "task-4294444" [ 1098.436842] env[62814]: _type = "Task" [ 1098.436842] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.446081] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294444, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.510842] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294441, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073206} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.511312] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1098.512357] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d9befe4-63bb-4c0f-9cc6-c2e5a39a2956 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.536698] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 201cf7be-f81f-4eba-ae65-a370ad8e48c3/201cf7be-f81f-4eba-ae65-a370ad8e48c3.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1098.537023] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99541b1e-4b05-44ec-ae19-bf3a1a6a8483 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.558305] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1098.558305] env[62814]: value = "task-4294445" [ 1098.558305] env[62814]: _type = "Task" [ 1098.558305] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.567429] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294445, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.718171] env[62814]: DEBUG oslo_vmware.api [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Task: {'id': task-4294443, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154883} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.718171] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.718171] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1098.718171] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1098.718171] env[62814]: INFO nova.compute.manager [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1098.718171] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1098.718551] env[62814]: DEBUG nova.compute.manager [-] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1098.718551] env[62814]: DEBUG nova.network.neutron [-] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.904273] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.959s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1098.907111] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.086s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1098.909332] env[62814]: INFO nova.compute.claims [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1098.932827] env[62814]: INFO nova.scheduler.client.report [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Deleted allocations for instance b4ee4cee-3298-4955-8375-8ca8c04b2f9f [ 1098.950068] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294444, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486655} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.950068] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 08c20588-2988-4976-8631-53cf447b4877/08c20588-2988-4976-8631-53cf447b4877.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1098.950432] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1098.950432] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac58f7a7-f3cd-4189-aa7b-2021d71f4fbf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.959316] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1098.959316] env[62814]: value = "task-4294446" [ 1098.959316] env[62814]: _type = "Task" [ 1098.959316] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.968291] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294446, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.046899] env[62814]: DEBUG nova.compute.manager [req-1fde769a-f51e-40df-9405-1a3a42e789cd req-8e381135-6ea5-48ef-9478-4aa06229ff7f service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Received event network-vif-deleted-5dbe3e06-4f03-4ee9-a3b8-e3906c07663a {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1099.047047] env[62814]: INFO nova.compute.manager [req-1fde769a-f51e-40df-9405-1a3a42e789cd req-8e381135-6ea5-48ef-9478-4aa06229ff7f service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Neutron deleted interface 5dbe3e06-4f03-4ee9-a3b8-e3906c07663a; detaching it from the instance and deleting it from the info cache [ 1099.047293] env[62814]: DEBUG nova.network.neutron [req-1fde769a-f51e-40df-9405-1a3a42e789cd req-8e381135-6ea5-48ef-9478-4aa06229ff7f service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.069994] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294445, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.138173] env[62814]: DEBUG nova.compute.manager [req-13b663c2-a742-46a2-89a2-7c12528536c3 req-4c9614b9-bbd6-415d-8059-0e6d50465e60 service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Received event network-vif-deleted-a00402c1-ebc9-40c0-93b2-26b6fbab4a68 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1099.139420] env[62814]: INFO nova.compute.manager [req-13b663c2-a742-46a2-89a2-7c12528536c3 req-4c9614b9-bbd6-415d-8059-0e6d50465e60 service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Neutron deleted interface a00402c1-ebc9-40c0-93b2-26b6fbab4a68; detaching it from the instance and deleting it from the info cache [ 1099.139420] env[62814]: DEBUG nova.network.neutron [req-13b663c2-a742-46a2-89a2-7c12528536c3 req-4c9614b9-bbd6-415d-8059-0e6d50465e60 service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.444047] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a8c07828-9f5d-42a1-a961-ea42bbfac6e2 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "b4ee4cee-3298-4955-8375-8ca8c04b2f9f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.974s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1099.469179] env[62814]: DEBUG nova.network.neutron [-] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.470577] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294446, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069592} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.470827] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1099.471644] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3568dd68-d743-4cb5-92dd-bad851a5bdd7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.497319] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 08c20588-2988-4976-8631-53cf447b4877/08c20588-2988-4976-8631-53cf447b4877.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.497970] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9554c776-548c-49f0-bdb7-770a3ef07674 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.519638] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1099.519638] env[62814]: value = "task-4294447" [ 1099.519638] env[62814]: _type = "Task" [ 1099.519638] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.528244] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294447, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.530776] env[62814]: DEBUG nova.network.neutron [-] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.549591] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd2598a2-db7d-4a0d-8206-1d0a76e0afa0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.559297] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f9c89b-8445-4076-a083-2053ec7fa690 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.579831] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294445, 'name': ReconfigVM_Task, 'duration_secs': 0.550842} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.580156] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 201cf7be-f81f-4eba-ae65-a370ad8e48c3/201cf7be-f81f-4eba-ae65-a370ad8e48c3.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1099.580713] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa66113a-f3d7-41dc-abac-d5db259de904 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.587137] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1099.587137] env[62814]: value = "task-4294448" [ 1099.587137] env[62814]: _type = "Task" [ 1099.587137] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.598870] env[62814]: DEBUG nova.compute.manager [req-1fde769a-f51e-40df-9405-1a3a42e789cd req-8e381135-6ea5-48ef-9478-4aa06229ff7f service nova] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Detach interface failed, port_id=5dbe3e06-4f03-4ee9-a3b8-e3906c07663a, reason: Instance 4429ff0c-5353-4346-8bef-2c45d251f1ea could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1099.607328] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294448, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.642806] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-814cebd2-479c-48b2-acf4-8f41350b2775 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.652885] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92fe77ee-2601-4e26-baf2-11caa7c31a6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.683856] env[62814]: DEBUG nova.compute.manager [req-13b663c2-a742-46a2-89a2-7c12528536c3 req-4c9614b9-bbd6-415d-8059-0e6d50465e60 service nova] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Detach interface failed, port_id=a00402c1-ebc9-40c0-93b2-26b6fbab4a68, reason: Instance 59d21ef7-df97-49ac-9329-4c18df6dd087 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1099.976068] env[62814]: INFO nova.compute.manager [-] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Took 1.57 seconds to deallocate network for instance. [ 1100.031503] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294447, 'name': ReconfigVM_Task, 'duration_secs': 0.297563} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.034289] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 08c20588-2988-4976-8631-53cf447b4877/08c20588-2988-4976-8631-53cf447b4877.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1100.035272] env[62814]: INFO nova.compute.manager [-] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Took 1.32 seconds to deallocate network for instance. [ 1100.035571] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce7b4fd4-8bbc-4fc0-a05d-4f31a5b9a3a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.043801] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1100.043801] env[62814]: value = "task-4294449" [ 1100.043801] env[62814]: _type = "Task" [ 1100.043801] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.054194] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294449, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.111500] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294448, 'name': Rename_Task, 'duration_secs': 0.143837} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.112052] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1100.112052] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d69d0e7-c2ab-44c3-9f72-f78bb65c9498 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.119414] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1100.119414] env[62814]: value = "task-4294450" [ 1100.119414] env[62814]: _type = "Task" [ 1100.119414] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.127061] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294450, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.203180] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4633c931-bee2-4cb7-bec5-36807d88f108 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.210987] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ef535d-6a70-4dba-a904-bd7e41a252a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.250450] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca831875-430a-4d0c-86c2-b6b6a848d2bb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.262150] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2697853-cb1c-4855-bd25-ab488df1fe42 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.275614] env[62814]: DEBUG nova.compute.provider_tree [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.375598] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1100.375737] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1100.376026] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1100.376202] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1100.376683] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1100.379575] env[62814]: INFO nova.compute.manager [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Terminating instance [ 1100.483198] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1100.543082] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1100.553691] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294449, 'name': Rename_Task, 'duration_secs': 0.197464} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.553974] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1100.554241] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a18b6541-59a7-4512-82bb-fb9c2260c154 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.560614] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1100.560614] env[62814]: value = "task-4294451" [ 1100.560614] env[62814]: _type = "Task" [ 1100.560614] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.567910] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294451, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.629553] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294450, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.782288] env[62814]: DEBUG nova.scheduler.client.report [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1100.884208] env[62814]: DEBUG nova.compute.manager [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1100.884208] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1100.884677] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b911cf7-0141-44fc-8c45-bf391b3bb539 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.892347] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1100.892611] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ced6b6f5-0de9-4da0-afde-6676d5b10907 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.899153] env[62814]: DEBUG oslo_vmware.api [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1100.899153] env[62814]: value = "task-4294452" [ 1100.899153] env[62814]: _type = "Task" [ 1100.899153] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.907433] env[62814]: DEBUG oslo_vmware.api [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.070961] env[62814]: DEBUG oslo_vmware.api [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294451, 'name': PowerOnVM_Task, 'duration_secs': 0.446706} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.071296] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1101.071540] env[62814]: INFO nova.compute.manager [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Took 6.49 seconds to spawn the instance on the hypervisor. [ 1101.071743] env[62814]: DEBUG nova.compute.manager [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1101.072784] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b5fdfa-811f-4481-8918-5291849090a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.134571] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294450, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.287382] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.287943] env[62814]: DEBUG nova.compute.manager [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1101.290797] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.948s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.291540] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.293237] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.329s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1101.296033] env[62814]: INFO nova.compute.claims [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1101.322643] env[62814]: INFO nova.scheduler.client.report [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleted allocations for instance 9343301c-dfe9-41b0-b4a0-067af544d297 [ 1101.409294] env[62814]: DEBUG oslo_vmware.api [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294452, 'name': PowerOffVM_Task, 'duration_secs': 0.217255} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.409573] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1101.409742] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1101.409990] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0af0eb9-bec9-4bf3-93f1-a0869f441b78 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.490067] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1101.490617] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1101.490617] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Deleting the datastore file [datastore2] 1c86646f-6f38-4f8d-bea0-8a6b93680aba {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1101.490881] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b33c4d1f-8ec6-4031-ad72-f5a49ef29d4c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.497362] env[62814]: DEBUG oslo_vmware.api [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1101.497362] env[62814]: value = "task-4294454" [ 1101.497362] env[62814]: _type = "Task" [ 1101.497362] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.505393] env[62814]: DEBUG oslo_vmware.api [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294454, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.589951] env[62814]: INFO nova.compute.manager [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Took 33.63 seconds to build instance. [ 1101.632524] env[62814]: DEBUG oslo_vmware.api [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294450, 'name': PowerOnVM_Task, 'duration_secs': 1.047101} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.632678] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1101.632870] env[62814]: INFO nova.compute.manager [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Took 9.59 seconds to spawn the instance on the hypervisor. [ 1101.633183] env[62814]: DEBUG nova.compute.manager [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1101.633993] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cc7e991-a397-4360-b09f-f4df336e251e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.801294] env[62814]: DEBUG nova.compute.utils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1101.804611] env[62814]: DEBUG nova.compute.manager [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1101.804779] env[62814]: DEBUG nova.network.neutron [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1101.831811] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2ebf6172-ab23-43b4-86f7-6b66e0fab1ea tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "9343301c-dfe9-41b0-b4a0-067af544d297" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.709s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1101.862848] env[62814]: DEBUG nova.policy [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e246f49b0b84cd093549b6d6b384e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6e52480dd2c467790622901940cf385', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1102.008301] env[62814]: DEBUG oslo_vmware.api [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294454, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.302384} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.008721] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1102.011018] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1102.011018] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1102.011018] env[62814]: INFO nova.compute.manager [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1102.011018] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1102.011018] env[62814]: DEBUG nova.compute.manager [-] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1102.011018] env[62814]: DEBUG nova.network.neutron [-] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1102.092510] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c667fa33-36a6-414b-92f3-015583da08f2 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "08c20588-2988-4976-8631-53cf447b4877" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.147s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.154125] env[62814]: INFO nova.compute.manager [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Took 35.51 seconds to build instance. [ 1102.254816] env[62814]: DEBUG nova.network.neutron [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Successfully created port: 0f65290c-3126-4fa7-9407-e7ad7e8018e2 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1102.305935] env[62814]: DEBUG nova.compute.manager [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1102.657971] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3c127ee9-03c6-4bab-8efb-98f14e4b8a0a tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.030s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1102.666748] env[62814]: DEBUG nova.compute.manager [req-07cdc3db-bd67-46ff-820c-e8132fea9543 req-f1e5b59a-0fbf-4008-99fb-432d76e9395a service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Received event network-vif-deleted-76bd40fc-ecff-43a9-a0e9-f58b9500f9c5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1102.667424] env[62814]: INFO nova.compute.manager [req-07cdc3db-bd67-46ff-820c-e8132fea9543 req-f1e5b59a-0fbf-4008-99fb-432d76e9395a service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Neutron deleted interface 76bd40fc-ecff-43a9-a0e9-f58b9500f9c5; detaching it from the instance and deleting it from the info cache [ 1102.667424] env[62814]: DEBUG nova.network.neutron [req-07cdc3db-bd67-46ff-820c-e8132fea9543 req-f1e5b59a-0fbf-4008-99fb-432d76e9395a service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.679852] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304aea3c-fab5-42a8-a46e-a93e23b3592c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.691317] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa01f0b6-0e33-4bee-bf6f-5c4e02dde3ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.724698] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e08a90-866f-4515-ae90-833c8b3a8a10 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.733141] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0267ba2e-cee7-49a7-a665-998a849fe6ee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.746973] env[62814]: DEBUG nova.compute.provider_tree [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.882553] env[62814]: DEBUG nova.network.neutron [-] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.021403] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "eeb0255e-8c94-4564-af87-818cc8dee58e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1103.021403] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "eeb0255e-8c94-4564-af87-818cc8dee58e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1103.169758] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fae83961-971f-46f3-aca7-924bc392cf8c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.183929] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0273981-0647-44b7-8693-93c77e38cfc4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.220890] env[62814]: DEBUG nova.compute.manager [req-07cdc3db-bd67-46ff-820c-e8132fea9543 req-f1e5b59a-0fbf-4008-99fb-432d76e9395a service nova] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Detach interface failed, port_id=76bd40fc-ecff-43a9-a0e9-f58b9500f9c5, reason: Instance 1c86646f-6f38-4f8d-bea0-8a6b93680aba could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1103.246307] env[62814]: DEBUG nova.compute.manager [req-7dd92d0c-1354-47fd-8c9c-45e99c516490 req-3a26ee8b-185f-4ba8-b74b-69ad6a196c74 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Received event network-changed-b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1103.246505] env[62814]: DEBUG nova.compute.manager [req-7dd92d0c-1354-47fd-8c9c-45e99c516490 req-3a26ee8b-185f-4ba8-b74b-69ad6a196c74 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Refreshing instance network info cache due to event network-changed-b3b985f6-c18c-47da-a863-93d8e8a87f19. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1103.248579] env[62814]: DEBUG oslo_concurrency.lockutils [req-7dd92d0c-1354-47fd-8c9c-45e99c516490 req-3a26ee8b-185f-4ba8-b74b-69ad6a196c74 service nova] Acquiring lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.248579] env[62814]: DEBUG oslo_concurrency.lockutils [req-7dd92d0c-1354-47fd-8c9c-45e99c516490 req-3a26ee8b-185f-4ba8-b74b-69ad6a196c74 service nova] Acquired lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1103.248579] env[62814]: DEBUG nova.network.neutron [req-7dd92d0c-1354-47fd-8c9c-45e99c516490 req-3a26ee8b-185f-4ba8-b74b-69ad6a196c74 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Refreshing network info cache for port b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1103.249936] env[62814]: DEBUG nova.scheduler.client.report [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1103.316872] env[62814]: DEBUG nova.compute.manager [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1103.344636] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1103.344885] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.345050] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1103.345235] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.345385] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1103.345551] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1103.345780] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1103.345962] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1103.346185] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1103.346408] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1103.346619] env[62814]: DEBUG nova.virt.hardware [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1103.347557] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c53c41fa-3df9-4a1e-b3a9-1cc6b9f27ad9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.355652] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47af74a4-08ed-47fa-af5f-36bf5bc88537 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.384697] env[62814]: INFO nova.compute.manager [-] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Took 1.37 seconds to deallocate network for instance. [ 1103.526357] env[62814]: DEBUG nova.compute.manager [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1103.756604] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1103.757221] env[62814]: DEBUG nova.compute.manager [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1103.760994] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.451s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1103.761449] env[62814]: DEBUG nova.objects.instance [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'resources' on Instance uuid 9683894b-a300-4400-a1b9-db62478f42c5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.893913] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.032496] env[62814]: DEBUG nova.network.neutron [req-7dd92d0c-1354-47fd-8c9c-45e99c516490 req-3a26ee8b-185f-4ba8-b74b-69ad6a196c74 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updated VIF entry in instance network info cache for port b3b985f6-c18c-47da-a863-93d8e8a87f19. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1104.032974] env[62814]: DEBUG nova.network.neutron [req-7dd92d0c-1354-47fd-8c9c-45e99c516490 req-3a26ee8b-185f-4ba8-b74b-69ad6a196c74 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updating instance_info_cache with network_info: [{"id": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "address": "fa:16:3e:31:60:d0", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3b985f6-c1", "ovs_interfaceid": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.046764] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.064751] env[62814]: DEBUG nova.network.neutron [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Successfully updated port: 0f65290c-3126-4fa7-9407-e7ad7e8018e2 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1104.266052] env[62814]: DEBUG nova.compute.utils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1104.269620] env[62814]: DEBUG nova.compute.manager [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1104.269918] env[62814]: DEBUG nova.network.neutron [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1104.310144] env[62814]: DEBUG nova.policy [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ca28a465227468a9f21fa208f16ca7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ed3e3bdfb4d4646a8c9eef1582cde85', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1104.527972] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1104.528295] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1104.538357] env[62814]: DEBUG oslo_concurrency.lockutils [req-7dd92d0c-1354-47fd-8c9c-45e99c516490 req-3a26ee8b-185f-4ba8-b74b-69ad6a196c74 service nova] Releasing lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1104.568619] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.568619] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1104.568619] env[62814]: DEBUG nova.network.neutron [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1104.633906] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ad66da-a548-4c4f-bed7-1e437e037462 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.641939] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10e0a44a-ccbf-47fd-9988-82fdba17354c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.672619] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c72ce2-03ed-4203-9f13-f6b7c3331afb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.680425] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4d1932d-b660-4ac2-a62b-2082af6f4ebf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.693807] env[62814]: DEBUG nova.compute.provider_tree [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.773144] env[62814]: DEBUG nova.compute.manager [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1104.850540] env[62814]: DEBUG nova.network.neutron [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Successfully created port: a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1105.030553] env[62814]: DEBUG nova.compute.manager [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1105.105031] env[62814]: DEBUG nova.network.neutron [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1105.197169] env[62814]: DEBUG nova.scheduler.client.report [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1105.263929] env[62814]: DEBUG nova.network.neutron [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Updating instance_info_cache with network_info: [{"id": "0f65290c-3126-4fa7-9407-e7ad7e8018e2", "address": "fa:16:3e:72:31:f5", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f65290c-31", "ovs_interfaceid": "0f65290c-3126-4fa7-9407-e7ad7e8018e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.557841] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.655506] env[62814]: DEBUG nova.compute.manager [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Received event network-vif-plugged-0f65290c-3126-4fa7-9407-e7ad7e8018e2 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1105.657023] env[62814]: DEBUG oslo_concurrency.lockutils [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] Acquiring lock "8af98433-18eb-4a12-b18d-39b0a58821c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1105.657023] env[62814]: DEBUG oslo_concurrency.lockutils [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.657023] env[62814]: DEBUG oslo_concurrency.lockutils [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.657023] env[62814]: DEBUG nova.compute.manager [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] No waiting events found dispatching network-vif-plugged-0f65290c-3126-4fa7-9407-e7ad7e8018e2 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1105.657023] env[62814]: WARNING nova.compute.manager [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Received unexpected event network-vif-plugged-0f65290c-3126-4fa7-9407-e7ad7e8018e2 for instance with vm_state building and task_state spawning. [ 1105.657023] env[62814]: DEBUG nova.compute.manager [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Received event network-changed-0f65290c-3126-4fa7-9407-e7ad7e8018e2 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1105.657023] env[62814]: DEBUG nova.compute.manager [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Refreshing instance network info cache due to event network-changed-0f65290c-3126-4fa7-9407-e7ad7e8018e2. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1105.657344] env[62814]: DEBUG oslo_concurrency.lockutils [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] Acquiring lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.703744] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.942s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1105.707324] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.112s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1105.708203] env[62814]: INFO nova.compute.claims [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1105.736343] env[62814]: INFO nova.scheduler.client.report [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Deleted allocations for instance 9683894b-a300-4400-a1b9-db62478f42c5 [ 1105.764941] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1105.765293] env[62814]: DEBUG nova.compute.manager [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Instance network_info: |[{"id": "0f65290c-3126-4fa7-9407-e7ad7e8018e2", "address": "fa:16:3e:72:31:f5", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f65290c-31", "ovs_interfaceid": "0f65290c-3126-4fa7-9407-e7ad7e8018e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1105.765617] env[62814]: DEBUG oslo_concurrency.lockutils [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] Acquired lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1105.765810] env[62814]: DEBUG nova.network.neutron [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Refreshing network info cache for port 0f65290c-3126-4fa7-9407-e7ad7e8018e2 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1105.767850] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:72:31:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '51bac3c3-00ab-4a07-9e28-b3c951dee565', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0f65290c-3126-4fa7-9407-e7ad7e8018e2', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1105.776957] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1105.777675] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1105.778412] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0e5c72bf-72a2-438d-9fc1-6c11ad3f9fd9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.798413] env[62814]: DEBUG nova.compute.manager [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1105.807130] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1105.807130] env[62814]: value = "task-4294455" [ 1105.807130] env[62814]: _type = "Task" [ 1105.807130] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.815132] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294455, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.826025] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1105.826232] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1105.826390] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1105.826573] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1105.826723] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1105.826870] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1105.827088] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1105.827253] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1105.827423] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1105.827587] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1105.827760] env[62814]: DEBUG nova.virt.hardware [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1105.828583] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293fcf54-27c4-445a-b0d6-9e89a61167f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.836536] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82298df7-af69-4cd2-ad36-67dc534e1f05 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.214811] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.215234] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.244705] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62644aad-ba5f-4aa3-8dcf-6e909a740a2e tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "9683894b-a300-4400-a1b9-db62478f42c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.817s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1106.320798] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294455, 'name': CreateVM_Task, 'duration_secs': 0.33212} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.323377] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1106.324216] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.324330] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1106.324655] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1106.325252] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1b9b8c0-16f0-474b-9d49-06b7d245ff38 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.333415] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1106.333415] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527e8b61-5db0-1f11-c34c-74392d912c85" [ 1106.333415] env[62814]: _type = "Task" [ 1106.333415] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.344657] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527e8b61-5db0-1f11-c34c-74392d912c85, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.552877] env[62814]: DEBUG nova.network.neutron [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Updated VIF entry in instance network info cache for port 0f65290c-3126-4fa7-9407-e7ad7e8018e2. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1106.552877] env[62814]: DEBUG nova.network.neutron [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Updating instance_info_cache with network_info: [{"id": "0f65290c-3126-4fa7-9407-e7ad7e8018e2", "address": "fa:16:3e:72:31:f5", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f65290c-31", "ovs_interfaceid": "0f65290c-3126-4fa7-9407-e7ad7e8018e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.724368] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.724567] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.724723] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.724903] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.725089] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.725249] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.725384] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 1106.725525] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1106.774760] env[62814]: DEBUG nova.compute.manager [req-c732ca28-5ab0-4180-881e-b19e576b6a99 req-fc6c9f19-3c04-40be-a685-017a6298daf3 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received event network-vif-plugged-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1106.775110] env[62814]: DEBUG oslo_concurrency.lockutils [req-c732ca28-5ab0-4180-881e-b19e576b6a99 req-fc6c9f19-3c04-40be-a685-017a6298daf3 service nova] Acquiring lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1106.775332] env[62814]: DEBUG oslo_concurrency.lockutils [req-c732ca28-5ab0-4180-881e-b19e576b6a99 req-fc6c9f19-3c04-40be-a685-017a6298daf3 service nova] Lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1106.775525] env[62814]: DEBUG oslo_concurrency.lockutils [req-c732ca28-5ab0-4180-881e-b19e576b6a99 req-fc6c9f19-3c04-40be-a685-017a6298daf3 service nova] Lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1106.775714] env[62814]: DEBUG nova.compute.manager [req-c732ca28-5ab0-4180-881e-b19e576b6a99 req-fc6c9f19-3c04-40be-a685-017a6298daf3 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] No waiting events found dispatching network-vif-plugged-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1106.775882] env[62814]: WARNING nova.compute.manager [req-c732ca28-5ab0-4180-881e-b19e576b6a99 req-fc6c9f19-3c04-40be-a685-017a6298daf3 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received unexpected event network-vif-plugged-a7104ea3-596d-4d10-952c-ac861087e580 for instance with vm_state building and task_state spawning. [ 1106.843729] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527e8b61-5db0-1f11-c34c-74392d912c85, 'name': SearchDatastore_Task, 'duration_secs': 0.033035} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.844041] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1106.844277] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1106.844506] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1106.844649] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1106.844824] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1106.845084] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-65485f51-1989-471b-9993-e3d9bfd104c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.855749] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1106.856037] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1106.856855] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef211b4b-bf94-48ea-880c-bd9e237705df {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.864531] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1106.864531] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52deb618-74d9-3cec-9609-e3210e0993c4" [ 1106.864531] env[62814]: _type = "Task" [ 1106.864531] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.873854] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52deb618-74d9-3cec-9609-e3210e0993c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.038080] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e62fac-4b83-4abc-972d-4fd8b81396a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.046240] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0261d32a-e3fc-4fc7-a384-78b6b88e068c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.078493] env[62814]: DEBUG oslo_concurrency.lockutils [req-42956af0-d5ed-4a7b-879c-53b42b2a2953 req-e401bfbd-6505-408a-b6a8-a3bbf4084a44 service nova] Releasing lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1107.080020] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1543c56e-a985-4e70-b6d8-29edab7fa800 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.087356] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1442d0c4-1b78-4313-a697-7ed5504b2d19 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.102623] env[62814]: DEBUG nova.compute.provider_tree [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1107.144092] env[62814]: DEBUG nova.network.neutron [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Successfully updated port: a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1107.229515] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1107.377221] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52deb618-74d9-3cec-9609-e3210e0993c4, 'name': SearchDatastore_Task, 'duration_secs': 0.011481} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.378308] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-895db783-b58e-49d4-9164-d7886e19f493 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.384029] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1107.384029] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52689af9-7c78-b1e3-daf8-f9095e7250cb" [ 1107.384029] env[62814]: _type = "Task" [ 1107.384029] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.393310] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52689af9-7c78-b1e3-daf8-f9095e7250cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.610788] env[62814]: DEBUG nova.scheduler.client.report [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1107.648510] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.648594] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1107.648775] env[62814]: DEBUG nova.network.neutron [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1107.774317] env[62814]: DEBUG nova.compute.manager [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Received event network-changed-b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1107.774512] env[62814]: DEBUG nova.compute.manager [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Refreshing instance network info cache due to event network-changed-b3b985f6-c18c-47da-a863-93d8e8a87f19. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1107.774728] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Acquiring lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.774873] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Acquired lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1107.775044] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Refreshing network info cache for port b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1107.895467] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52689af9-7c78-b1e3-daf8-f9095e7250cb, 'name': SearchDatastore_Task, 'duration_secs': 0.030814} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.895752] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1107.895998] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 8af98433-18eb-4a12-b18d-39b0a58821c0/8af98433-18eb-4a12-b18d-39b0a58821c0.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1107.896339] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-acc5120c-8626-49fd-8829-af11aa3f3fed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.905033] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1107.905033] env[62814]: value = "task-4294456" [ 1107.905033] env[62814]: _type = "Task" [ 1107.905033] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.913272] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294456, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.100360] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "08c20588-2988-4976-8631-53cf447b4877" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.100858] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "08c20588-2988-4976-8631-53cf447b4877" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.100858] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "08c20588-2988-4976-8631-53cf447b4877-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.101011] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "08c20588-2988-4976-8631-53cf447b4877-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.101198] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "08c20588-2988-4976-8631-53cf447b4877-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1108.107967] env[62814]: INFO nova.compute.manager [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Terminating instance [ 1108.117350] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.410s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1108.117835] env[62814]: DEBUG nova.compute.manager [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1108.121770] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.649s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.121770] env[62814]: DEBUG nova.objects.instance [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lazy-loading 'resources' on Instance uuid 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.202747] env[62814]: DEBUG nova.network.neutron [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1108.416012] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294456, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.418469] env[62814]: DEBUG nova.network.neutron [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating instance_info_cache with network_info: [{"id": "a7104ea3-596d-4d10-952c-ac861087e580", "address": "fa:16:3e:4a:60:cd", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7104ea3-59", "ovs_interfaceid": "a7104ea3-596d-4d10-952c-ac861087e580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.596793] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updated VIF entry in instance network info cache for port b3b985f6-c18c-47da-a863-93d8e8a87f19. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1108.597254] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updating instance_info_cache with network_info: [{"id": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "address": "fa:16:3e:31:60:d0", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.147", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3b985f6-c1", "ovs_interfaceid": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.616133] env[62814]: DEBUG nova.compute.manager [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1108.616397] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1108.617348] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aeb8fdb-0928-4c15-8150-fb5e55ece2d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.630154] env[62814]: DEBUG nova.compute.utils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1108.632787] env[62814]: DEBUG nova.compute.manager [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1108.632966] env[62814]: DEBUG nova.network.neutron [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1108.641663] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1108.642049] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a524889-1fa8-46c7-bb98-e2bdb563decd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.650686] env[62814]: DEBUG oslo_vmware.api [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1108.650686] env[62814]: value = "task-4294457" [ 1108.650686] env[62814]: _type = "Task" [ 1108.650686] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.663560] env[62814]: DEBUG oslo_vmware.api [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294457, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.712501] env[62814]: DEBUG nova.policy [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd05b1b390e0489d91f4f386f4c6fe41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65a886f76f404ce5a6d0841241d8f120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1108.774421] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1108.774421] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1108.919916] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294456, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.889581} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.920585] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 8af98433-18eb-4a12-b18d-39b0a58821c0/8af98433-18eb-4a12-b18d-39b0a58821c0.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1108.920828] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1108.927668] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1108.928243] env[62814]: DEBUG nova.compute.manager [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Instance network_info: |[{"id": "a7104ea3-596d-4d10-952c-ac861087e580", "address": "fa:16:3e:4a:60:cd", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7104ea3-59", "ovs_interfaceid": "a7104ea3-596d-4d10-952c-ac861087e580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1108.928374] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c1efc891-fb2a-49db-87bc-876ea1b0927a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.931932] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:60:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7104ea3-596d-4d10-952c-ac861087e580', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1108.942048] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating folder: Project (8ed3e3bdfb4d4646a8c9eef1582cde85). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1108.942048] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ecda9fe-888c-49f5-9a16-6cbcd5b761e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.946580] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1108.946580] env[62814]: value = "task-4294458" [ 1108.946580] env[62814]: _type = "Task" [ 1108.946580] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.952920] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Created folder: Project (8ed3e3bdfb4d4646a8c9eef1582cde85) in parent group-v845547. [ 1108.953126] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating folder: Instances. Parent ref: group-v845821. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1108.956621] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3dd5abe-cf00-4f02-a5d5-d526eebd30c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.958423] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294458, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.968901] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Created folder: Instances in parent group-v845821. [ 1108.969163] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1108.969362] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1108.969578] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2233387b-9d47-4f39-a947-3e6f46c7a447 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.994504] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1108.994504] env[62814]: value = "task-4294461" [ 1108.994504] env[62814]: _type = "Task" [ 1108.994504] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.003799] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294461, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.055202] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1308dd-20f4-44c2-9491-cd5c0f9e7280 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.063449] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d76e7e8-9841-4b0d-9e78-b9ad9c61165b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.095912] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef31a3e8-009d-4055-9b9b-37bc69b6ef52 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.099447] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Releasing lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1109.099712] env[62814]: DEBUG nova.compute.manager [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Received event network-changed-8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1109.100660] env[62814]: DEBUG nova.compute.manager [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Refreshing instance network info cache due to event network-changed-8fb80dc8-60f2-403a-99f7-4c77d3431369. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1109.100660] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Acquiring lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.100660] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Acquired lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1109.100660] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Refreshing network info cache for port 8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1109.105313] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcb20fd-1a6c-43d2-bb91-b2056fa8ceaa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.120125] env[62814]: DEBUG nova.compute.provider_tree [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1109.137783] env[62814]: DEBUG nova.compute.manager [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1109.160451] env[62814]: DEBUG oslo_vmware.api [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294457, 'name': PowerOffVM_Task, 'duration_secs': 0.30304} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.160564] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1109.160706] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1109.161333] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-519318f1-6b3d-4168-97db-b85a3cdf9155 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.242030] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1109.242030] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1109.242030] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Deleting the datastore file [datastore2] 08c20588-2988-4976-8631-53cf447b4877 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1109.242393] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21893941-2637-41bd-a889-f313632bbc16 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.249375] env[62814]: DEBUG oslo_vmware.api [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1109.249375] env[62814]: value = "task-4294463" [ 1109.249375] env[62814]: _type = "Task" [ 1109.249375] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.262016] env[62814]: DEBUG oslo_vmware.api [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294463, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.262329] env[62814]: DEBUG nova.network.neutron [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Successfully created port: 3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1109.279322] env[62814]: DEBUG nova.compute.manager [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1109.455529] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294458, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063118} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.455800] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1109.457029] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8830d0-68ec-48c8-8504-a89b481e36ac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.482568] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 8af98433-18eb-4a12-b18d-39b0a58821c0/8af98433-18eb-4a12-b18d-39b0a58821c0.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1109.482891] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ee8b8f5-bcef-40d0-a744-b228a4005be3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.513951] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294461, 'name': CreateVM_Task, 'duration_secs': 0.33429} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.515264] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1109.516708] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1109.516708] env[62814]: value = "task-4294464" [ 1109.516708] env[62814]: _type = "Task" [ 1109.516708] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.516708] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.516708] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1109.516957] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1109.517064] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8a139be-f787-40e6-9766-d1784f0ac606 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.526877] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1109.526877] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a6ca52-a5bc-a853-4d8f-ac61eb81a8aa" [ 1109.526877] env[62814]: _type = "Task" [ 1109.526877] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.531476] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294464, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.540971] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a6ca52-a5bc-a853-4d8f-ac61eb81a8aa, 'name': SearchDatastore_Task, 'duration_secs': 0.009932} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.541322] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1109.541577] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1109.541834] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.542013] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1109.542209] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1109.542552] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ddfaad54-284e-46db-856a-46bd2f368c87 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.550866] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1109.551083] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1109.551831] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1652a48a-7934-4128-b05e-4bdd7a5aaa40 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.556967] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1109.556967] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523eba6b-fd3e-df6a-24af-31c062f0f1e3" [ 1109.556967] env[62814]: _type = "Task" [ 1109.556967] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.564857] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523eba6b-fd3e-df6a-24af-31c062f0f1e3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.623159] env[62814]: DEBUG nova.scheduler.client.report [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1109.763259] env[62814]: DEBUG oslo_vmware.api [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294463, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163396} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.763680] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1109.763989] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1109.764324] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1109.764625] env[62814]: INFO nova.compute.manager [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 08c20588-2988-4976-8631-53cf447b4877] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1109.764909] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1109.765139] env[62814]: DEBUG nova.compute.manager [-] [instance: 08c20588-2988-4976-8631-53cf447b4877] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1109.765244] env[62814]: DEBUG nova.network.neutron [-] [instance: 08c20588-2988-4976-8631-53cf447b4877] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1109.810701] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1110.027354] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294464, 'name': ReconfigVM_Task, 'duration_secs': 0.263267} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.027653] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 8af98433-18eb-4a12-b18d-39b0a58821c0/8af98433-18eb-4a12-b18d-39b0a58821c0.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.028284] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7af5ec0b-d5ed-498e-8dd2-00083a9c6f31 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.038019] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1110.038019] env[62814]: value = "task-4294465" [ 1110.038019] env[62814]: _type = "Task" [ 1110.038019] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.045016] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294465, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.066975] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523eba6b-fd3e-df6a-24af-31c062f0f1e3, 'name': SearchDatastore_Task, 'duration_secs': 0.008405} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.067806] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0636e0dc-370e-4e99-9a3c-a629a98dc98a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.073658] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1110.073658] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522b65ef-a317-85b4-bc46-6117e36e6acd" [ 1110.073658] env[62814]: _type = "Task" [ 1110.073658] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.084581] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522b65ef-a317-85b4-bc46-6117e36e6acd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.128389] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.007s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1110.132685] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.342s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1110.134724] env[62814]: INFO nova.compute.claims [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1110.157620] env[62814]: DEBUG nova.compute.manager [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1110.166360] env[62814]: INFO nova.scheduler.client.report [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Deleted allocations for instance 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3 [ 1110.170534] env[62814]: DEBUG nova.compute.manager [req-e069048d-f5f3-42a1-92e2-08d7359934c1 req-9f6ae8b4-e5a6-4bf9-9929-59d82b96d8f8 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Received event network-changed-b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1110.170744] env[62814]: DEBUG nova.compute.manager [req-e069048d-f5f3-42a1-92e2-08d7359934c1 req-9f6ae8b4-e5a6-4bf9-9929-59d82b96d8f8 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Refreshing instance network info cache due to event network-changed-b3b985f6-c18c-47da-a863-93d8e8a87f19. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1110.170968] env[62814]: DEBUG oslo_concurrency.lockutils [req-e069048d-f5f3-42a1-92e2-08d7359934c1 req-9f6ae8b4-e5a6-4bf9-9929-59d82b96d8f8 service nova] Acquiring lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.172050] env[62814]: DEBUG oslo_concurrency.lockutils [req-e069048d-f5f3-42a1-92e2-08d7359934c1 req-9f6ae8b4-e5a6-4bf9-9929-59d82b96d8f8 service nova] Acquired lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1110.172050] env[62814]: DEBUG nova.network.neutron [req-e069048d-f5f3-42a1-92e2-08d7359934c1 req-9f6ae8b4-e5a6-4bf9-9929-59d82b96d8f8 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Refreshing network info cache for port b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1110.188232] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1110.188551] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1110.188732] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1110.188940] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1110.189115] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1110.189287] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1110.189528] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1110.189734] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1110.190051] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1110.190317] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1110.190519] env[62814]: DEBUG nova.virt.hardware [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1110.191728] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c204d2-b109-4d74-9ca3-cca104c17a1e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.200943] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1a1afea-a5b0-4f71-85a1-45ec3eb1c05c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.220878] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Updated VIF entry in instance network info cache for port 8fb80dc8-60f2-403a-99f7-4c77d3431369. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1110.221227] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Updating instance_info_cache with network_info: [{"id": "8fb80dc8-60f2-403a-99f7-4c77d3431369", "address": "fa:16:3e:4d:02:44", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8fb80dc8-60", "ovs_interfaceid": "8fb80dc8-60f2-403a-99f7-4c77d3431369", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.372522] env[62814]: DEBUG nova.compute.manager [req-853c9678-bdf7-46f6-b8e4-c288220c8059 req-3b7bf430-592a-4969-89d2-bfce4a48a2db service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Received event network-vif-deleted-8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1110.372722] env[62814]: INFO nova.compute.manager [req-853c9678-bdf7-46f6-b8e4-c288220c8059 req-3b7bf430-592a-4969-89d2-bfce4a48a2db service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Neutron deleted interface 8fb80dc8-60f2-403a-99f7-4c77d3431369; detaching it from the instance and deleting it from the info cache [ 1110.372894] env[62814]: DEBUG nova.network.neutron [req-853c9678-bdf7-46f6-b8e4-c288220c8059 req-3b7bf430-592a-4969-89d2-bfce4a48a2db service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.546877] env[62814]: DEBUG nova.network.neutron [-] [instance: 08c20588-2988-4976-8631-53cf447b4877] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.548474] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294465, 'name': Rename_Task, 'duration_secs': 0.139717} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.548756] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1110.549361] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff30fdc2-9af5-49fa-aabf-ff9da00778b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.556808] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1110.556808] env[62814]: value = "task-4294466" [ 1110.556808] env[62814]: _type = "Task" [ 1110.556808] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.566273] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294466, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.586043] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522b65ef-a317-85b4-bc46-6117e36e6acd, 'name': SearchDatastore_Task, 'duration_secs': 0.010384} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.586043] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1110.586043] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1f13051e-e51b-4981-9445-d5420c3c1818/1f13051e-e51b-4981-9445-d5420c3c1818.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1110.586043] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9f94482c-9eec-40f4-8823-135fc3b3eab5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.591845] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1110.591845] env[62814]: value = "task-4294467" [ 1110.591845] env[62814]: _type = "Task" [ 1110.591845] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.600364] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294467, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.681915] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f73eef2c-2970-4c23-b5b6-8bfd1621085b tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "19b53797-6fc9-4bb0-ab3f-9d2941b11cb3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.884s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1110.725485] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Releasing lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1110.725485] env[62814]: DEBUG nova.compute.manager [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Received event network-changed-8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1110.725485] env[62814]: DEBUG nova.compute.manager [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Refreshing instance network info cache due to event network-changed-8fb80dc8-60f2-403a-99f7-4c77d3431369. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1110.725485] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Acquiring lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1110.725485] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Acquired lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1110.725485] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Refreshing network info cache for port 8fb80dc8-60f2-403a-99f7-4c77d3431369 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1110.875353] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ce8446c6-2986-4fd3-9b1f-9b9d35cf0e5e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.888618] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8da694e-7bb1-468a-b2f9-0cd1d4dd9dda {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.923908] env[62814]: DEBUG nova.compute.manager [req-853c9678-bdf7-46f6-b8e4-c288220c8059 req-3b7bf430-592a-4969-89d2-bfce4a48a2db service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Detach interface failed, port_id=8fb80dc8-60f2-403a-99f7-4c77d3431369, reason: Instance 08c20588-2988-4976-8631-53cf447b4877 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1111.004934] env[62814]: DEBUG nova.network.neutron [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Successfully updated port: 3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1111.007943] env[62814]: DEBUG nova.network.neutron [req-e069048d-f5f3-42a1-92e2-08d7359934c1 req-9f6ae8b4-e5a6-4bf9-9929-59d82b96d8f8 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updated VIF entry in instance network info cache for port b3b985f6-c18c-47da-a863-93d8e8a87f19. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1111.008299] env[62814]: DEBUG nova.network.neutron [req-e069048d-f5f3-42a1-92e2-08d7359934c1 req-9f6ae8b4-e5a6-4bf9-9929-59d82b96d8f8 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updating instance_info_cache with network_info: [{"id": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "address": "fa:16:3e:31:60:d0", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3b985f6-c1", "ovs_interfaceid": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.050094] env[62814]: INFO nova.compute.manager [-] [instance: 08c20588-2988-4976-8631-53cf447b4877] Took 1.28 seconds to deallocate network for instance. [ 1111.067897] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294466, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.107135] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294467, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510291} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.107135] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1f13051e-e51b-4981-9445-d5420c3c1818/1f13051e-e51b-4981-9445-d5420c3c1818.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1111.107135] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1111.107135] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a0e4ffdc-50cb-4304-9fde-9495409a7599 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.113922] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1111.113922] env[62814]: value = "task-4294468" [ 1111.113922] env[62814]: _type = "Task" [ 1111.113922] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.122976] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294468, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.259181] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1111.373064] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 08c20588-2988-4976-8631-53cf447b4877] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1111.468017] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53df9dca-7bf0-4fa7-b78a-60d38456dcd3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.475156] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9772c683-b58b-44b3-ae8c-86f98b194a1e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.509164] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "refresh_cache-f00d414e-1dfd-43cf-9245-7cbeea8850b3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.509334] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "refresh_cache-f00d414e-1dfd-43cf-9245-7cbeea8850b3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1111.509495] env[62814]: DEBUG nova.network.neutron [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1111.511913] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-955e9684-f778-4045-87f2-892909528f83 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.515387] env[62814]: DEBUG oslo_concurrency.lockutils [req-e069048d-f5f3-42a1-92e2-08d7359934c1 req-9f6ae8b4-e5a6-4bf9-9929-59d82b96d8f8 service nova] Releasing lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1111.521738] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44227bc4-5734-4e9f-a572-2dad18e8ac58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.539487] env[62814]: DEBUG nova.compute.provider_tree [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1111.557125] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.569184] env[62814]: DEBUG oslo_vmware.api [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294466, 'name': PowerOnVM_Task, 'duration_secs': 0.725819} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.569184] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1111.569184] env[62814]: INFO nova.compute.manager [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Took 8.25 seconds to spawn the instance on the hypervisor. [ 1111.569307] env[62814]: DEBUG nova.compute.manager [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1111.570190] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50d8ba7-8e8e-4509-8714-31696b6296bb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.624131] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294468, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06558} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.624416] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1111.625623] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcad98c3-356e-431e-b434-04780d125164 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.648796] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 1f13051e-e51b-4981-9445-d5420c3c1818/1f13051e-e51b-4981-9445-d5420c3c1818.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1111.649089] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c5586b6-416f-43b9-bdd5-5237ad31dbe9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.663517] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "e727eeb1-c5d0-4591-80bb-31746bf976a2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.663767] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "e727eeb1-c5d0-4591-80bb-31746bf976a2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.664573] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "e727eeb1-c5d0-4591-80bb-31746bf976a2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1111.664573] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "e727eeb1-c5d0-4591-80bb-31746bf976a2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1111.664573] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "e727eeb1-c5d0-4591-80bb-31746bf976a2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1111.666447] env[62814]: INFO nova.compute.manager [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Terminating instance [ 1111.674303] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1111.674303] env[62814]: value = "task-4294469" [ 1111.674303] env[62814]: _type = "Task" [ 1111.674303] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.683510] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294469, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.876643] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Releasing lock "refresh_cache-08c20588-2988-4976-8631-53cf447b4877" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1111.876643] env[62814]: DEBUG nova.compute.manager [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received event network-changed-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1111.876775] env[62814]: DEBUG nova.compute.manager [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Refreshing instance network info cache due to event network-changed-a7104ea3-596d-4d10-952c-ac861087e580. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1111.876887] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Acquiring lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.877071] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Acquired lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1111.877205] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Refreshing network info cache for port a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1112.044114] env[62814]: DEBUG nova.scheduler.client.report [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1112.048065] env[62814]: DEBUG nova.network.neutron [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1112.094875] env[62814]: INFO nova.compute.manager [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Took 37.29 seconds to build instance. [ 1112.170409] env[62814]: DEBUG nova.compute.manager [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1112.170673] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1112.171650] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2dc877-659d-462b-a336-16b2db435504 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.182933] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294469, 'name': ReconfigVM_Task, 'duration_secs': 0.336885} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.185115] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 1f13051e-e51b-4981-9445-d5420c3c1818/1f13051e-e51b-4981-9445-d5420c3c1818.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.185801] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1112.186681] env[62814]: DEBUG nova.network.neutron [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Updating instance_info_cache with network_info: [{"id": "3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5", "address": "fa:16:3e:cb:8e:9b", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d97fdd4-3a", "ovs_interfaceid": "3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.188026] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-844fd33f-b888-46b1-aa35-52d152e78db2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.189553] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-229eaa4e-a024-426f-8e5e-add62bf46973 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.196269] env[62814]: DEBUG oslo_vmware.api [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1112.196269] env[62814]: value = "task-4294471" [ 1112.196269] env[62814]: _type = "Task" [ 1112.196269] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.197494] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1112.197494] env[62814]: value = "task-4294470" [ 1112.197494] env[62814]: _type = "Task" [ 1112.197494] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.210711] env[62814]: DEBUG oslo_vmware.api [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294471, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.214381] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294470, 'name': Rename_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.242578] env[62814]: DEBUG nova.compute.manager [req-650724bd-cc58-4c31-9bbc-e8683a861166 req-3b3ca74a-8bc2-46e9-a5a9-628e27ffee70 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Received event network-changed-b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1112.242691] env[62814]: DEBUG nova.compute.manager [req-650724bd-cc58-4c31-9bbc-e8683a861166 req-3b3ca74a-8bc2-46e9-a5a9-628e27ffee70 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Refreshing instance network info cache due to event network-changed-b3b985f6-c18c-47da-a863-93d8e8a87f19. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1112.242878] env[62814]: DEBUG oslo_concurrency.lockutils [req-650724bd-cc58-4c31-9bbc-e8683a861166 req-3b3ca74a-8bc2-46e9-a5a9-628e27ffee70 service nova] Acquiring lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.242963] env[62814]: DEBUG oslo_concurrency.lockutils [req-650724bd-cc58-4c31-9bbc-e8683a861166 req-3b3ca74a-8bc2-46e9-a5a9-628e27ffee70 service nova] Acquired lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1112.243134] env[62814]: DEBUG nova.network.neutron [req-650724bd-cc58-4c31-9bbc-e8683a861166 req-3b3ca74a-8bc2-46e9-a5a9-628e27ffee70 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Refreshing network info cache for port b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1112.551830] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.552790] env[62814]: DEBUG nova.compute.manager [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1112.555458] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 21.392s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.597702] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c6e2744c-d008-4901-b2e6-49109f0ad713 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.807s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.615187] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updated VIF entry in instance network info cache for port a7104ea3-596d-4d10-952c-ac861087e580. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1112.615548] env[62814]: DEBUG nova.network.neutron [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating instance_info_cache with network_info: [{"id": "a7104ea3-596d-4d10-952c-ac861087e580", "address": "fa:16:3e:4a:60:cd", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7104ea3-59", "ovs_interfaceid": "a7104ea3-596d-4d10-952c-ac861087e580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.692607] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "refresh_cache-f00d414e-1dfd-43cf-9245-7cbeea8850b3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1112.693025] env[62814]: DEBUG nova.compute.manager [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Instance network_info: |[{"id": "3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5", "address": "fa:16:3e:cb:8e:9b", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d97fdd4-3a", "ovs_interfaceid": "3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1112.693491] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cb:8e:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7e0240aa-a694-48fc-a0f9-6f2d3e71aa12', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1112.701465] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1112.702782] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1112.703967] env[62814]: DEBUG nova.compute.manager [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Received event network-vif-plugged-3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1112.704180] env[62814]: DEBUG oslo_concurrency.lockutils [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] Acquiring lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.704388] env[62814]: DEBUG oslo_concurrency.lockutils [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.704552] env[62814]: DEBUG oslo_concurrency.lockutils [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1112.704715] env[62814]: DEBUG nova.compute.manager [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] No waiting events found dispatching network-vif-plugged-3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1112.704877] env[62814]: WARNING nova.compute.manager [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Received unexpected event network-vif-plugged-3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5 for instance with vm_state building and task_state spawning. [ 1112.705050] env[62814]: DEBUG nova.compute.manager [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Received event network-changed-3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1112.705269] env[62814]: DEBUG nova.compute.manager [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Refreshing instance network info cache due to event network-changed-3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1112.705614] env[62814]: DEBUG oslo_concurrency.lockutils [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] Acquiring lock "refresh_cache-f00d414e-1dfd-43cf-9245-7cbeea8850b3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.705803] env[62814]: DEBUG oslo_concurrency.lockutils [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] Acquired lock "refresh_cache-f00d414e-1dfd-43cf-9245-7cbeea8850b3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1112.705970] env[62814]: DEBUG nova.network.neutron [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Refreshing network info cache for port 3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1112.710877] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ad0e829-2473-4ef3-bef6-4a99c2dbd212 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.737250] env[62814]: DEBUG oslo_vmware.api [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294471, 'name': PowerOffVM_Task, 'duration_secs': 0.210594} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.741173] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1112.741306] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1112.741958] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1112.741958] env[62814]: value = "task-4294472" [ 1112.741958] env[62814]: _type = "Task" [ 1112.741958] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.742137] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294470, 'name': Rename_Task, 'duration_secs': 0.150151} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.742550] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e6ceeab-29b5-4477-b8b9-ba34f3451e17 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.743934] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1112.747183] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be0ec652-aa81-417f-acc1-52280a76b6d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.758215] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294472, 'name': CreateVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.759567] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1112.759567] env[62814]: value = "task-4294474" [ 1112.759567] env[62814]: _type = "Task" [ 1112.759567] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.769815] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294474, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.774348] env[62814]: DEBUG oslo_concurrency.lockutils [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "8af98433-18eb-4a12-b18d-39b0a58821c0" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1112.774584] env[62814]: DEBUG oslo_concurrency.lockutils [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1112.775217] env[62814]: INFO nova.compute.manager [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Shelving [ 1112.828618] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1112.828998] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1112.829105] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Deleting the datastore file [datastore2] e727eeb1-c5d0-4591-80bb-31746bf976a2 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1112.829868] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42a73e27-c560-4530-95e3-c4084005b5f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.835659] env[62814]: DEBUG oslo_vmware.api [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for the task: (returnval){ [ 1112.835659] env[62814]: value = "task-4294475" [ 1112.835659] env[62814]: _type = "Task" [ 1112.835659] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.844027] env[62814]: DEBUG oslo_vmware.api [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294475, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.999723] env[62814]: DEBUG nova.network.neutron [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Updated VIF entry in instance network info cache for port 3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1113.000150] env[62814]: DEBUG nova.network.neutron [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Updating instance_info_cache with network_info: [{"id": "3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5", "address": "fa:16:3e:cb:8e:9b", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d97fdd4-3a", "ovs_interfaceid": "3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.012101] env[62814]: DEBUG nova.network.neutron [req-650724bd-cc58-4c31-9bbc-e8683a861166 req-3b3ca74a-8bc2-46e9-a5a9-628e27ffee70 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updated VIF entry in instance network info cache for port b3b985f6-c18c-47da-a863-93d8e8a87f19. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1113.012692] env[62814]: DEBUG nova.network.neutron [req-650724bd-cc58-4c31-9bbc-e8683a861166 req-3b3ca74a-8bc2-46e9-a5a9-628e27ffee70 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updating instance_info_cache with network_info: [{"id": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "address": "fa:16:3e:31:60:d0", "network": {"id": "b6a57152-3574-4cad-a8d4-b17daeaeaa78", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationTestJSON-2022214679-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "67cc9d091e094518a5ba7b78708698fe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c883fb98-d172-4510-8cf4-07aafdf771af", "external-id": "nsx-vlan-transportzone-570", "segmentation_id": 570, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3b985f6-c1", "ovs_interfaceid": "b3b985f6-c18c-47da-a863-93d8e8a87f19", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.060754] env[62814]: DEBUG nova.compute.utils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1113.065909] env[62814]: INFO nova.compute.claims [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1113.068479] env[62814]: DEBUG nova.compute.manager [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1113.068997] env[62814]: DEBUG nova.network.neutron [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1113.117424] env[62814]: DEBUG nova.policy [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aafa735af2694aa486b41728d304b81b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5fe4b0c679ba4104a781dc54aad98ef5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1113.119729] env[62814]: DEBUG oslo_concurrency.lockutils [req-6da69843-cdb1-4477-ba51-a76eaa574374 req-adef631a-4190-4073-a308-a426cfb7b6ee service nova] Releasing lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.164254] env[62814]: INFO nova.compute.manager [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Rebuilding instance [ 1113.226230] env[62814]: DEBUG nova.compute.manager [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1113.230025] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57946a0d-dc98-42ea-94b4-749402d0fb1d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.253517] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294472, 'name': CreateVM_Task, 'duration_secs': 0.314627} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.253730] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1113.254464] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.254705] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1113.254989] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1113.255265] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47631e25-c7dc-4b48-9cae-8b468a0f0789 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.260349] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1113.260349] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52888523-bb96-57f9-4a9a-6d04cc167d8d" [ 1113.260349] env[62814]: _type = "Task" [ 1113.260349] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.272433] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294474, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.276132] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52888523-bb96-57f9-4a9a-6d04cc167d8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.347226] env[62814]: DEBUG oslo_vmware.api [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Task: {'id': task-4294475, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.209537} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.347626] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1113.351018] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1113.351018] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1113.351018] env[62814]: INFO nova.compute.manager [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Took 1.18 seconds to destroy the instance on the hypervisor. [ 1113.351018] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1113.351018] env[62814]: DEBUG nova.compute.manager [-] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1113.351018] env[62814]: DEBUG nova.network.neutron [-] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1113.504782] env[62814]: DEBUG oslo_concurrency.lockutils [req-af378cee-ca25-4415-8570-62c716022c63 req-46761f7f-5774-4776-b4bd-ce83c7a29787 service nova] Releasing lock "refresh_cache-f00d414e-1dfd-43cf-9245-7cbeea8850b3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.515862] env[62814]: DEBUG oslo_concurrency.lockutils [req-650724bd-cc58-4c31-9bbc-e8683a861166 req-3b3ca74a-8bc2-46e9-a5a9-628e27ffee70 service nova] Releasing lock "refresh_cache-0c6c598f-ef5a-4e91-b811-cd3d8f072647" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.572877] env[62814]: DEBUG nova.compute.manager [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1113.574511] env[62814]: INFO nova.compute.resource_tracker [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating resource usage from migration ef6d923a-0f7c-438d-bc3f-962f524c5f40 [ 1113.773513] env[62814]: DEBUG oslo_vmware.api [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294474, 'name': PowerOnVM_Task, 'duration_secs': 0.602043} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.777880] env[62814]: DEBUG nova.network.neutron [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Successfully created port: 5c6118e7-498d-4a79-a493-8029099cd533 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1113.780305] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1113.780542] env[62814]: INFO nova.compute.manager [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Took 7.98 seconds to spawn the instance on the hypervisor. [ 1113.780731] env[62814]: DEBUG nova.compute.manager [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1113.781036] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52888523-bb96-57f9-4a9a-6d04cc167d8d, 'name': SearchDatastore_Task, 'duration_secs': 0.011814} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.786224] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a26371-fb54-42ad-81e2-799dc8030573 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.790744] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1113.790971] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1113.791243] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.791390] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1113.791574] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1113.792362] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1113.793623] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-66da1524-b6df-4b99-864d-bf01062fff23 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.794619] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2df0893d-1333-4ac9-808a-ff477734745f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.806248] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1113.806248] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1113.807027] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1113.807027] env[62814]: value = "task-4294476" [ 1113.807027] env[62814]: _type = "Task" [ 1113.807027] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.807610] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfd72036-8cd5-4524-b173-2c6c4c109ceb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.819083] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1113.819083] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e7e950-85a1-916c-629e-6a46d499e84e" [ 1113.819083] env[62814]: _type = "Task" [ 1113.819083] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.822509] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294476, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.833297] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e7e950-85a1-916c-629e-6a46d499e84e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.957710] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c00156-98bf-4230-a85c-9424e31dc6d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.965808] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35f1739f-c8cf-453f-b2cf-d0a2ed118921 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.998049] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fe888fe-fb7d-42d3-a99b-104f1a499e1f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.006599] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4251e4-d623-4965-95eb-99b4a38ab414 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.022859] env[62814]: DEBUG nova.compute.provider_tree [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1114.159564] env[62814]: DEBUG nova.network.neutron [-] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.249085] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.249996] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13b0b0af-a63b-42b7-b442-fae9ecdb2ec5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.259104] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for the task: (returnval){ [ 1114.259104] env[62814]: value = "task-4294477" [ 1114.259104] env[62814]: _type = "Task" [ 1114.259104] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.267924] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294477, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.322488] env[62814]: INFO nova.compute.manager [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Took 35.39 seconds to build instance. [ 1114.329902] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294476, 'name': PowerOffVM_Task, 'duration_secs': 0.214773} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.331461] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1114.331461] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0acea16-cba4-4518-95de-f992d52e2a36 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.337572] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e7e950-85a1-916c-629e-6a46d499e84e, 'name': SearchDatastore_Task, 'duration_secs': 0.013386} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.338500] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-164842ac-751e-4ac9-abc2-7cb9ddef33cc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.358737] env[62814]: DEBUG nova.compute.manager [req-f33d7795-6d68-42b5-95bd-91b89cc66e2b req-1ad433cc-17b7-4ce9-ad53-97319bf6a108 service nova] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Received event network-vif-deleted-9f8d050a-def9-4d6c-a47d-f3ac8d07b4bb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1114.360639] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ac58ac-450e-4b81-b0b3-da4d86d06c7c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.365429] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1114.365429] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523a4664-b13b-254e-147e-11b90df46aea" [ 1114.365429] env[62814]: _type = "Task" [ 1114.365429] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.381256] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523a4664-b13b-254e-147e-11b90df46aea, 'name': SearchDatastore_Task, 'duration_secs': 0.01373} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.381938] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1114.382375] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] f00d414e-1dfd-43cf-9245-7cbeea8850b3/f00d414e-1dfd-43cf-9245-7cbeea8850b3.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1114.382951] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-642f703b-8f55-48fd-a549-4eabef846964 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.393018] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1114.393018] env[62814]: value = "task-4294478" [ 1114.393018] env[62814]: _type = "Task" [ 1114.393018] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.400861] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294478, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.529213] env[62814]: DEBUG nova.scheduler.client.report [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1114.585507] env[62814]: DEBUG nova.compute.manager [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1114.614054] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1114.614306] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1114.614534] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1114.614768] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1114.614929] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1114.615203] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1114.615487] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1114.615675] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1114.615851] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1114.616033] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1114.616216] env[62814]: DEBUG nova.virt.hardware [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1114.617430] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8c5d9e-5a0a-4dc4-a157-5eec2c057114 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.625257] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b995995-b5d4-4fe1-9704-fe2351a4fbc5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.662241] env[62814]: INFO nova.compute.manager [-] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Took 1.31 seconds to deallocate network for instance. [ 1114.784724] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294477, 'name': PowerOffVM_Task, 'duration_secs': 0.163707} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.785045] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1114.785814] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.786118] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fccd0651-d314-446d-aeb5-b99e5858b9b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.796181] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for the task: (returnval){ [ 1114.796181] env[62814]: value = "task-4294479" [ 1114.796181] env[62814]: _type = "Task" [ 1114.796181] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.813706] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1114.814092] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1114.814390] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845786', 'volume_id': '6f2eddd0-4334-4e22-b195-b7deaa539235', 'name': 'volume-6f2eddd0-4334-4e22-b195-b7deaa539235', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd5df1d0e-55e6-452e-882a-053a83250b39', 'attached_at': '', 'detached_at': '', 'volume_id': '6f2eddd0-4334-4e22-b195-b7deaa539235', 'serial': '6f2eddd0-4334-4e22-b195-b7deaa539235'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1114.815726] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6403bb7-a75b-467e-81d0-0180eba712df {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.847047] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33674f0-fee7-4313-b914-f06d086621fc tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.920s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1114.848348] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7608b75f-7280-482e-b141-f4394b4cdfb1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.860269] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc38264-958e-44dc-9f88-b96b55ae8ff2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.879780] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1114.880189] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-b2866218-6aec-4e1a-a391-4e3c6500d488 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.882820] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f60ccc-4fd9-4e17-a1db-02ba478f9885 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.891451] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1114.891451] env[62814]: value = "task-4294480" [ 1114.891451] env[62814]: _type = "Task" [ 1114.891451] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.902303] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] The volume has not been displaced from its original location: [datastore2] volume-6f2eddd0-4334-4e22-b195-b7deaa539235/volume-6f2eddd0-4334-4e22-b195-b7deaa539235.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1114.907834] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Reconfiguring VM instance instance-0000005b to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1114.914852] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01ffcef9-76c0-4ed5-b970-f3bf2b9740a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.929690] env[62814]: DEBUG nova.compute.manager [req-0fdfa775-d5e9-40e3-bc84-9c336a0b5dfb req-7bab65a2-92f6-43cf-9197-9fccc7b0b246 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received event network-changed-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1114.930131] env[62814]: DEBUG nova.compute.manager [req-0fdfa775-d5e9-40e3-bc84-9c336a0b5dfb req-7bab65a2-92f6-43cf-9197-9fccc7b0b246 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Refreshing instance network info cache due to event network-changed-a7104ea3-596d-4d10-952c-ac861087e580. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1114.930361] env[62814]: DEBUG oslo_concurrency.lockutils [req-0fdfa775-d5e9-40e3-bc84-9c336a0b5dfb req-7bab65a2-92f6-43cf-9197-9fccc7b0b246 service nova] Acquiring lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.930497] env[62814]: DEBUG oslo_concurrency.lockutils [req-0fdfa775-d5e9-40e3-bc84-9c336a0b5dfb req-7bab65a2-92f6-43cf-9197-9fccc7b0b246 service nova] Acquired lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1114.930651] env[62814]: DEBUG nova.network.neutron [req-0fdfa775-d5e9-40e3-bc84-9c336a0b5dfb req-7bab65a2-92f6-43cf-9197-9fccc7b0b246 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Refreshing network info cache for port a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1114.939350] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294478, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.941796] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for the task: (returnval){ [ 1114.941796] env[62814]: value = "task-4294481" [ 1114.941796] env[62814]: _type = "Task" [ 1114.941796] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.952461] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294481, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.035903] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.480s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1115.038030] env[62814]: INFO nova.compute.manager [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Migrating [ 1115.043995] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.234s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1115.046160] env[62814]: INFO nova.compute.claims [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1115.169203] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1115.422031] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294478, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.699898} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.422337] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] f00d414e-1dfd-43cf-9245-7cbeea8850b3/f00d414e-1dfd-43cf-9245-7cbeea8850b3.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1115.422765] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1115.423137] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294480, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.425493] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ef2d9f39-0eb0-4e84-81d3-752c986234aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.430415] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1115.430415] env[62814]: value = "task-4294482" [ 1115.430415] env[62814]: _type = "Task" [ 1115.430415] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.447469] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294482, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.457885] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294481, 'name': ReconfigVM_Task, 'duration_secs': 0.241761} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.458068] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Reconfigured VM instance instance-0000005b to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1115.463092] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75cec966-451d-4502-a38b-d63e46b7d124 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.479241] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for the task: (returnval){ [ 1115.479241] env[62814]: value = "task-4294483" [ 1115.479241] env[62814]: _type = "Task" [ 1115.479241] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.490397] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294483, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.512886] env[62814]: DEBUG nova.network.neutron [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Successfully updated port: 5c6118e7-498d-4a79-a493-8029099cd533 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1115.561361] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.561548] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1115.561778] env[62814]: DEBUG nova.network.neutron [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1115.812293] env[62814]: DEBUG nova.network.neutron [req-0fdfa775-d5e9-40e3-bc84-9c336a0b5dfb req-7bab65a2-92f6-43cf-9197-9fccc7b0b246 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updated VIF entry in instance network info cache for port a7104ea3-596d-4d10-952c-ac861087e580. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1115.812293] env[62814]: DEBUG nova.network.neutron [req-0fdfa775-d5e9-40e3-bc84-9c336a0b5dfb req-7bab65a2-92f6-43cf-9197-9fccc7b0b246 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating instance_info_cache with network_info: [{"id": "a7104ea3-596d-4d10-952c-ac861087e580", "address": "fa:16:3e:4a:60:cd", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7104ea3-59", "ovs_interfaceid": "a7104ea3-596d-4d10-952c-ac861087e580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.913205] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294480, 'name': CreateSnapshot_Task, 'duration_secs': 0.999476} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.914584] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1115.914584] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c44288-1720-44da-90a1-7380c1e8cd24 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.940764] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294482, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132307} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.941133] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1115.941982] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7027c672-0518-464a-9b10-418f964f1c5b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.966078] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] f00d414e-1dfd-43cf-9245-7cbeea8850b3/f00d414e-1dfd-43cf-9245-7cbeea8850b3.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1115.967205] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b67eaeb-a9bb-4d2a-89dc-d6123124221b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.991790] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294483, 'name': ReconfigVM_Task, 'duration_secs': 0.159053} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.992687] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845786', 'volume_id': '6f2eddd0-4334-4e22-b195-b7deaa539235', 'name': 'volume-6f2eddd0-4334-4e22-b195-b7deaa539235', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'd5df1d0e-55e6-452e-882a-053a83250b39', 'attached_at': '', 'detached_at': '', 'volume_id': '6f2eddd0-4334-4e22-b195-b7deaa539235', 'serial': '6f2eddd0-4334-4e22-b195-b7deaa539235'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1115.993348] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1115.993348] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1115.993348] env[62814]: value = "task-4294484" [ 1115.993348] env[62814]: _type = "Task" [ 1115.993348] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.993990] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae1f7b90-a218-455d-9a4f-256380b6c361 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.007635] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1116.007895] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-15e345b4-7fbc-40dd-a94b-c2eacf9c0556 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.015169] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquiring lock "refresh_cache-e75c8039-9ff8-45f2-8770-894f2307cc14" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.015398] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquired lock "refresh_cache-e75c8039-9ff8-45f2-8770-894f2307cc14" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1116.015901] env[62814]: DEBUG nova.network.neutron [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1116.088561] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1116.088873] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1116.089101] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Deleting the datastore file [datastore2] d5df1d0e-55e6-452e-882a-053a83250b39 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1116.089411] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4995995a-ddd6-4150-bde2-4e012e0e118c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.097129] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for the task: (returnval){ [ 1116.097129] env[62814]: value = "task-4294486" [ 1116.097129] env[62814]: _type = "Task" [ 1116.097129] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.109894] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294486, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.315467] env[62814]: DEBUG oslo_concurrency.lockutils [req-0fdfa775-d5e9-40e3-bc84-9c336a0b5dfb req-7bab65a2-92f6-43cf-9197-9fccc7b0b246 service nova] Releasing lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1116.360506] env[62814]: DEBUG nova.network.neutron [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance_info_cache with network_info: [{"id": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "address": "fa:16:3e:35:63:2b", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1881a6c-56", "ovs_interfaceid": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.398355] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be61ee67-f32a-4ee1-81f5-bba8e7d07fa1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.406828] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573f620b-071b-48d1-88ed-5067ad4daa0b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.444816] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1116.445925] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-cdc1c740-a99c-4656-9160-aa106e48c51e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.449755] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a49951-f19a-4408-aa6a-6b6dca6b0608 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.459444] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7722f42f-f57f-4e24-9e26-60ca3ad40ad6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.463308] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1116.463308] env[62814]: value = "task-4294487" [ 1116.463308] env[62814]: _type = "Task" [ 1116.463308] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1116.474501] env[62814]: DEBUG nova.compute.provider_tree [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1116.480883] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294487, 'name': CloneVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.507777] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294484, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.570919] env[62814]: DEBUG nova.network.neutron [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1116.608136] env[62814]: DEBUG oslo_vmware.api [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Task: {'id': task-4294486, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134995} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.608136] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1116.608345] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1116.608572] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1116.676310] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1116.676652] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10ca017a-cab9-4e27-bc69-e766816108f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.686048] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e137532-b43a-46a0-9548-ae72733dd2f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.720368] env[62814]: ERROR nova.compute.manager [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Failed to detach volume 6f2eddd0-4334-4e22-b195-b7deaa539235 from /dev/sda: nova.exception.InstanceNotFound: Instance d5df1d0e-55e6-452e-882a-053a83250b39 could not be found. [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Traceback (most recent call last): [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 4185, in _do_rebuild_instance [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self.driver.rebuild(**kwargs) [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] raise NotImplementedError() [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] NotImplementedError [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] During handling of the above exception, another exception occurred: [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Traceback (most recent call last): [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 3608, in _detach_root_volume [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self.driver.detach_volume(context, old_connection_info, [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] return self._volumeops.detach_volume(connection_info, instance) [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self._detach_volume_vmdk(connection_info, instance) [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] stable_ref.fetch_moref(session) [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] nova.exception.InstanceNotFound: Instance d5df1d0e-55e6-452e-882a-053a83250b39 could not be found. [ 1116.720368] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] [ 1116.762067] env[62814]: DEBUG nova.network.neutron [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Updating instance_info_cache with network_info: [{"id": "5c6118e7-498d-4a79-a493-8029099cd533", "address": "fa:16:3e:5a:3f:82", "network": {"id": "650c56bc-f5c1-4a14-83ff-d24d5e8bae71", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1692253125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fe4b0c679ba4104a781dc54aad98ef5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e839c46-1ae9-43b7-9518-8f18f48100dd", "external-id": "nsx-vlan-transportzone-666", "segmentation_id": 666, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c6118e7-49", "ovs_interfaceid": "5c6118e7-498d-4a79-a493-8029099cd533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.866349] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1116.873717] env[62814]: DEBUG nova.compute.utils [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Build of instance d5df1d0e-55e6-452e-882a-053a83250b39 aborted: Failed to rebuild volume backed instance. {{(pid=62814) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 1116.876218] env[62814]: ERROR nova.compute.manager [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance d5df1d0e-55e6-452e-882a-053a83250b39 aborted: Failed to rebuild volume backed instance. [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Traceback (most recent call last): [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 4185, in _do_rebuild_instance [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self.driver.rebuild(**kwargs) [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/driver.py", line 497, in rebuild [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] raise NotImplementedError() [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] NotImplementedError [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] During handling of the above exception, another exception occurred: [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Traceback (most recent call last): [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 3643, in _rebuild_volume_backed_instance [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self._detach_root_volume(context, instance, root_bdm) [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 3622, in _detach_root_volume [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] with excutils.save_and_reraise_exception(): [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self.force_reraise() [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] raise self.value [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 3608, in _detach_root_volume [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self.driver.detach_volume(context, old_connection_info, [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 561, in detach_volume [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] return self._volumeops.detach_volume(connection_info, instance) [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self._detach_volume_vmdk(connection_info, instance) [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] stable_ref.fetch_moref(session) [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] nova.exception.InstanceNotFound: Instance d5df1d0e-55e6-452e-882a-053a83250b39 could not be found. [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] During handling of the above exception, another exception occurred: [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Traceback (most recent call last): [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 11502, in _error_out_instance_on_exception [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] yield [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 3911, in rebuild_instance [ 1116.876218] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self._do_rebuild_instance_with_claim( [ 1116.877425] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 3997, in _do_rebuild_instance_with_claim [ 1116.877425] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self._do_rebuild_instance( [ 1116.877425] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 4189, in _do_rebuild_instance [ 1116.877425] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self._rebuild_default_impl(**kwargs) [ 1116.877425] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 3766, in _rebuild_default_impl [ 1116.877425] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] self._rebuild_volume_backed_instance( [ 1116.877425] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] File "/opt/stack/nova/nova/compute/manager.py", line 3658, in _rebuild_volume_backed_instance [ 1116.877425] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] raise exception.BuildAbortException( [ 1116.877425] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] nova.exception.BuildAbortException: Build of instance d5df1d0e-55e6-452e-882a-053a83250b39 aborted: Failed to rebuild volume backed instance. [ 1116.877425] env[62814]: ERROR nova.compute.manager [instance: d5df1d0e-55e6-452e-882a-053a83250b39] [ 1116.975534] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294487, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.977483] env[62814]: DEBUG nova.scheduler.client.report [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1117.008039] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294484, 'name': ReconfigVM_Task, 'duration_secs': 0.659959} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1117.008261] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Reconfigured VM instance instance-00000063 to attach disk [datastore2] f00d414e-1dfd-43cf-9245-7cbeea8850b3/f00d414e-1dfd-43cf-9245-7cbeea8850b3.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1117.008891] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-caf2dbaa-43bc-4cb1-8b5d-235219adc84c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.017044] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1117.017044] env[62814]: value = "task-4294488" [ 1117.017044] env[62814]: _type = "Task" [ 1117.017044] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.025592] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294488, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.058263] env[62814]: DEBUG nova.compute.manager [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Received event network-vif-plugged-5c6118e7-498d-4a79-a493-8029099cd533 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1117.058263] env[62814]: DEBUG oslo_concurrency.lockutils [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] Acquiring lock "e75c8039-9ff8-45f2-8770-894f2307cc14-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1117.058263] env[62814]: DEBUG oslo_concurrency.lockutils [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] Lock "e75c8039-9ff8-45f2-8770-894f2307cc14-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1117.058263] env[62814]: DEBUG oslo_concurrency.lockutils [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] Lock "e75c8039-9ff8-45f2-8770-894f2307cc14-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.058263] env[62814]: DEBUG nova.compute.manager [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] No waiting events found dispatching network-vif-plugged-5c6118e7-498d-4a79-a493-8029099cd533 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1117.058263] env[62814]: WARNING nova.compute.manager [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Received unexpected event network-vif-plugged-5c6118e7-498d-4a79-a493-8029099cd533 for instance with vm_state building and task_state spawning. [ 1117.058721] env[62814]: DEBUG nova.compute.manager [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Received event network-changed-5c6118e7-498d-4a79-a493-8029099cd533 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1117.058721] env[62814]: DEBUG nova.compute.manager [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Refreshing instance network info cache due to event network-changed-5c6118e7-498d-4a79-a493-8029099cd533. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1117.058721] env[62814]: DEBUG oslo_concurrency.lockutils [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] Acquiring lock "refresh_cache-e75c8039-9ff8-45f2-8770-894f2307cc14" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1117.265795] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Releasing lock "refresh_cache-e75c8039-9ff8-45f2-8770-894f2307cc14" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1117.266160] env[62814]: DEBUG nova.compute.manager [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Instance network_info: |[{"id": "5c6118e7-498d-4a79-a493-8029099cd533", "address": "fa:16:3e:5a:3f:82", "network": {"id": "650c56bc-f5c1-4a14-83ff-d24d5e8bae71", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1692253125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fe4b0c679ba4104a781dc54aad98ef5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e839c46-1ae9-43b7-9518-8f18f48100dd", "external-id": "nsx-vlan-transportzone-666", "segmentation_id": 666, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c6118e7-49", "ovs_interfaceid": "5c6118e7-498d-4a79-a493-8029099cd533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1117.266468] env[62814]: DEBUG oslo_concurrency.lockutils [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] Acquired lock "refresh_cache-e75c8039-9ff8-45f2-8770-894f2307cc14" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1117.266651] env[62814]: DEBUG nova.network.neutron [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Refreshing network info cache for port 5c6118e7-498d-4a79-a493-8029099cd533 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1117.268049] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:3f:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e839c46-1ae9-43b7-9518-8f18f48100dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c6118e7-498d-4a79-a493-8029099cd533', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1117.275133] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Creating folder: Project (5fe4b0c679ba4104a781dc54aad98ef5). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1117.278115] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-97d9e166-f501-43c6-b7d8-d0173c3d4199 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.289291] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Created folder: Project (5fe4b0c679ba4104a781dc54aad98ef5) in parent group-v845547. [ 1117.289523] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Creating folder: Instances. Parent ref: group-v845827. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1117.289804] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-85c87ca0-e78e-41fc-b563-328dfeda52cf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.998074] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.954s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1117.998540] env[62814]: DEBUG nova.compute.manager [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1118.001086] env[62814]: DEBUG oslo_concurrency.lockutils [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1118.001301] env[62814]: DEBUG oslo_concurrency.lockutils [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1118.001466] env[62814]: INFO nova.compute.manager [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Shelving [ 1118.004383] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Created folder: Instances in parent group-v845827. [ 1118.004601] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1118.008203] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.657s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1118.008414] env[62814]: DEBUG nova.objects.instance [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lazy-loading 'resources' on Instance uuid 44ea319c-6ea0-456a-bee6-42133a25d8c1 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.012192] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1118.013092] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cdd66b66-2bf2-47b3-a900-535f276bf189 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.033558] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294487, 'name': CloneVM_Task, 'duration_secs': 1.252086} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.037532] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Created linked-clone VM from snapshot [ 1118.037781] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1118.037781] env[62814]: value = "task-4294491" [ 1118.037781] env[62814]: _type = "Task" [ 1118.037781] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.038028] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294488, 'name': Rename_Task, 'duration_secs': 0.169688} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.038944] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ef5084-7f29-4cc0-b578-ac4bcb896121 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.041723] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1118.044946] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2d578651-9389-4cfa-890c-26aa3716e1f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.051216] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Uploading image f65ac645-8a26-45ab-b54f-1c8edc8064bd {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1118.058749] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294491, 'name': CreateVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.059304] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1118.059304] env[62814]: value = "task-4294492" [ 1118.059304] env[62814]: _type = "Task" [ 1118.059304] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.070399] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294492, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.082117] env[62814]: DEBUG oslo_vmware.rw_handles [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1118.082117] env[62814]: value = "vm-845826" [ 1118.082117] env[62814]: _type = "VirtualMachine" [ 1118.082117] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1118.082488] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a75c9443-ddf9-4a96-bd72-1573616c910a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.089603] env[62814]: DEBUG oslo_vmware.rw_handles [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lease: (returnval){ [ 1118.089603] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef57c0-afe0-81f4-a959-4bf3b899693e" [ 1118.089603] env[62814]: _type = "HttpNfcLease" [ 1118.089603] env[62814]: } obtained for exporting VM: (result){ [ 1118.089603] env[62814]: value = "vm-845826" [ 1118.089603] env[62814]: _type = "VirtualMachine" [ 1118.089603] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1118.089992] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the lease: (returnval){ [ 1118.089992] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef57c0-afe0-81f4-a959-4bf3b899693e" [ 1118.089992] env[62814]: _type = "HttpNfcLease" [ 1118.089992] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1118.098301] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1118.098301] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef57c0-afe0-81f4-a959-4bf3b899693e" [ 1118.098301] env[62814]: _type = "HttpNfcLease" [ 1118.098301] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1118.260518] env[62814]: DEBUG nova.network.neutron [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Updated VIF entry in instance network info cache for port 5c6118e7-498d-4a79-a493-8029099cd533. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1118.261131] env[62814]: DEBUG nova.network.neutron [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Updating instance_info_cache with network_info: [{"id": "5c6118e7-498d-4a79-a493-8029099cd533", "address": "fa:16:3e:5a:3f:82", "network": {"id": "650c56bc-f5c1-4a14-83ff-d24d5e8bae71", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1692253125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5fe4b0c679ba4104a781dc54aad98ef5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e839c46-1ae9-43b7-9518-8f18f48100dd", "external-id": "nsx-vlan-transportzone-666", "segmentation_id": 666, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c6118e7-49", "ovs_interfaceid": "5c6118e7-498d-4a79-a493-8029099cd533", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.512971] env[62814]: DEBUG nova.compute.utils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1118.517486] env[62814]: DEBUG nova.compute.manager [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1118.517707] env[62814]: DEBUG nova.network.neutron [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1118.552481] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294491, 'name': CreateVM_Task, 'duration_secs': 0.309868} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.552764] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1118.553390] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.553596] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1118.553935] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1118.554197] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a07d463-a1b7-4d98-9800-5d61fc4f6624 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.558692] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for the task: (returnval){ [ 1118.558692] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209a9d4-7c33-02e2-81a0-7ee23799a0ef" [ 1118.558692] env[62814]: _type = "Task" [ 1118.558692] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.570432] env[62814]: DEBUG nova.policy [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5779eaa60ec44b2b80660e521b598eb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10018060297c409d82935ad626cc54c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1118.580676] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209a9d4-7c33-02e2-81a0-7ee23799a0ef, 'name': SearchDatastore_Task, 'duration_secs': 0.010003} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.580943] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294492, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.583561] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1118.583849] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1118.584092] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.584237] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1118.584410] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1118.585308] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3366aa08-c797-4f2f-8659-8b88b36d7cbc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.595354] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1118.595630] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1118.598694] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dabb3a5-3847-40ae-9efb-fe84f5b91097 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.603316] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1118.603316] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef57c0-afe0-81f4-a959-4bf3b899693e" [ 1118.603316] env[62814]: _type = "HttpNfcLease" [ 1118.603316] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1118.605272] env[62814]: DEBUG oslo_vmware.rw_handles [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1118.605272] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ef57c0-afe0-81f4-a959-4bf3b899693e" [ 1118.605272] env[62814]: _type = "HttpNfcLease" [ 1118.605272] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1118.605272] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa61ec61-39bf-426c-901e-0de71e2b57f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.608685] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for the task: (returnval){ [ 1118.608685] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529a18e9-b28f-2267-6bcb-0788db89e863" [ 1118.608685] env[62814]: _type = "Task" [ 1118.608685] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.616701] env[62814]: DEBUG oslo_vmware.rw_handles [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52149d50-d11d-050a-d310-d42ff28d11a3/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1118.616701] env[62814]: DEBUG oslo_vmware.rw_handles [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52149d50-d11d-050a-d310-d42ff28d11a3/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1118.626890] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529a18e9-b28f-2267-6bcb-0788db89e863, 'name': SearchDatastore_Task, 'duration_secs': 0.013757} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.685182] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-531a5fe7-c4a4-4acf-a7ea-c40b120acb38 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.695107] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for the task: (returnval){ [ 1118.695107] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5288b3e5-e190-e959-db65-82089fc1d793" [ 1118.695107] env[62814]: _type = "Task" [ 1118.695107] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.703535] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5288b3e5-e190-e959-db65-82089fc1d793, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.724652] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-51b9a492-31e9-4c64-a998-5ae1ea77ea7a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.768407] env[62814]: DEBUG oslo_concurrency.lockutils [req-af3e3036-fd6c-4dcf-9d81-762e0b38c68e req-136fc9b3-fec4-4bd5-8377-7986bb2edf87 service nova] Releasing lock "refresh_cache-e75c8039-9ff8-45f2-8770-894f2307cc14" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1118.928074] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda8d3c2-8776-45a8-a69b-2439b534c0f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.937024] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849dead6-6d99-409c-848a-1b62491f663e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.971812] env[62814]: DEBUG nova.network.neutron [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Successfully created port: 1c6b72d7-d148-4280-981c-9bfbf768bd3b {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1118.975470] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fcdedf8-0811-4f28-9137-ebdcab4f4b13 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.983592] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c671c3-275c-4b2a-b531-3269f3586497 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.998419] env[62814]: DEBUG nova.compute.provider_tree [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1119.020706] env[62814]: DEBUG nova.compute.manager [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1119.023312] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1119.024137] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20113872-918a-477f-8c35-729dd601cc73 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.027993] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee66a098-546a-40a0-8b30-d78021eb49a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.047020] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance '7e849b23-226d-4a1d-a85c-cf96964ce034' progress to 0 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1119.053241] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1119.053241] env[62814]: value = "task-4294494" [ 1119.053241] env[62814]: _type = "Task" [ 1119.053241] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.062379] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294494, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.073356] env[62814]: DEBUG oslo_vmware.api [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294492, 'name': PowerOnVM_Task, 'duration_secs': 0.686735} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.073637] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1119.073956] env[62814]: INFO nova.compute.manager [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Took 8.92 seconds to spawn the instance on the hypervisor. [ 1119.074099] env[62814]: DEBUG nova.compute.manager [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1119.074780] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886a74ee-a581-43c6-bec8-8f742260c036 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.205894] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5288b3e5-e190-e959-db65-82089fc1d793, 'name': SearchDatastore_Task, 'duration_secs': 0.02849} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.206353] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1119.206638] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] e75c8039-9ff8-45f2-8770-894f2307cc14/e75c8039-9ff8-45f2-8770-894f2307cc14.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1119.207311] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-deb00508-bd39-47b7-91d3-d423a6a6bf4a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.214087] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for the task: (returnval){ [ 1119.214087] env[62814]: value = "task-4294495" [ 1119.214087] env[62814]: _type = "Task" [ 1119.214087] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.223195] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294495, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.502216] env[62814]: DEBUG nova.scheduler.client.report [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1119.524819] env[62814]: DEBUG oslo_concurrency.lockutils [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1119.559463] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1119.559807] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-058aaaf9-dfca-486a-80c0-edc93961410a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.575328] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294494, 'name': PowerOffVM_Task, 'duration_secs': 0.296801} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.576135] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1119.576741] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1119.576741] env[62814]: value = "task-4294496" [ 1119.576741] env[62814]: _type = "Task" [ 1119.576741] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.577563] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49c2474f-dde7-43a3-ba90-b54db6d73ebe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.612306] env[62814]: INFO nova.compute.manager [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Took 32.04 seconds to build instance. [ 1119.618316] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e26eecef-4782-4747-b304-0cd664e69109 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.621863] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294496, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.725648] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294495, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.011430] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.003s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.013859] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.531s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.014069] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.015804] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.473s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.016035] env[62814]: DEBUG nova.objects.instance [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lazy-loading 'resources' on Instance uuid 4429ff0c-5353-4346-8bef-2c45d251f1ea {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.037026] env[62814]: DEBUG nova.compute.manager [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1120.042188] env[62814]: INFO nova.scheduler.client.report [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleted allocations for instance 59d21ef7-df97-49ac-9329-4c18df6dd087 [ 1120.045507] env[62814]: INFO nova.scheduler.client.report [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Deleted allocations for instance 44ea319c-6ea0-456a-bee6-42133a25d8c1 [ 1120.071945] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1120.072351] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1120.072571] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1120.072777] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1120.073018] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1120.073865] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1120.073865] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1120.074010] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1120.074243] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1120.074420] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1120.074597] env[62814]: DEBUG nova.virt.hardware [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1120.075492] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8427659f-c88f-4af3-8341-60c22090218d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.088118] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb0ebd3-7c6e-4eb0-8d0a-0ab3c8d96e69 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.098427] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294496, 'name': PowerOffVM_Task, 'duration_secs': 0.47719} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.107399] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1120.107700] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance '7e849b23-226d-4a1d-a85c-cf96964ce034' progress to 17 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1120.116739] env[62814]: DEBUG nova.compute.manager [req-575e3b8e-0450-4cb6-a37d-208805ab96ee req-c886d29a-3f68-4cb3-bf45-a0ea4ff1658b service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Received event network-changed-3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1120.116950] env[62814]: DEBUG nova.compute.manager [req-575e3b8e-0450-4cb6-a37d-208805ab96ee req-c886d29a-3f68-4cb3-bf45-a0ea4ff1658b service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Refreshing instance network info cache due to event network-changed-3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1120.117203] env[62814]: DEBUG oslo_concurrency.lockutils [req-575e3b8e-0450-4cb6-a37d-208805ab96ee req-c886d29a-3f68-4cb3-bf45-a0ea4ff1658b service nova] Acquiring lock "refresh_cache-f00d414e-1dfd-43cf-9245-7cbeea8850b3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1120.117314] env[62814]: DEBUG oslo_concurrency.lockutils [req-575e3b8e-0450-4cb6-a37d-208805ab96ee req-c886d29a-3f68-4cb3-bf45-a0ea4ff1658b service nova] Acquired lock "refresh_cache-f00d414e-1dfd-43cf-9245-7cbeea8850b3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1120.117463] env[62814]: DEBUG nova.network.neutron [req-575e3b8e-0450-4cb6-a37d-208805ab96ee req-c886d29a-3f68-4cb3-bf45-a0ea4ff1658b service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Refreshing network info cache for port 3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1120.123138] env[62814]: DEBUG oslo_concurrency.lockutils [None req-324e4344-adce-4bc2-aaab-c5dca779f5c1 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.565s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.135075] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1120.135832] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-dc229188-9f3e-44e1-bb4d-d5ba90e562a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.144358] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1120.144358] env[62814]: value = "task-4294497" [ 1120.144358] env[62814]: _type = "Task" [ 1120.144358] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.153831] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294497, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.226644] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294495, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.673521} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.226945] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] e75c8039-9ff8-45f2-8770-894f2307cc14/e75c8039-9ff8-45f2-8770-894f2307cc14.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1120.227347] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1120.227460] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5bf85449-7f89-4fb2-b9e1-13fcefc6cb3d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.234663] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for the task: (returnval){ [ 1120.234663] env[62814]: value = "task-4294498" [ 1120.234663] env[62814]: _type = "Task" [ 1120.234663] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.243930] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294498, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.563588] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c3ab581a-fb21-4da1-ae95-6422852fab78 tempest-ServerRescueTestJSONUnderV235-897685578 tempest-ServerRescueTestJSONUnderV235-897685578-project-member] Lock "44ea319c-6ea0-456a-bee6-42133a25d8c1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.873s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.564991] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eed4c37b-9fde-48a2-ab29-b87f89638148 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "59d21ef7-df97-49ac-9329-4c18df6dd087" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.796s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.614401] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1120.614655] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1120.614860] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1120.615066] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1120.615213] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1120.615356] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1120.615558] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1120.615713] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1120.615877] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1120.616044] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1120.616328] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1120.627873] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe9c424b-8971-4507-8cb1-dd5d206534e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.652866] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1120.652866] env[62814]: value = "task-4294499" [ 1120.652866] env[62814]: _type = "Task" [ 1120.652866] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.660083] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294497, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.668140] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.745553] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294498, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.122182} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.745848] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1120.746692] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10303605-dfc1-4ece-b3c7-4403f9081a17 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.776314] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] e75c8039-9ff8-45f2-8770-894f2307cc14/e75c8039-9ff8-45f2-8770-894f2307cc14.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1120.776670] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6016f22f-3fa0-4e0b-bec0-333c1dda8e65 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.803158] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for the task: (returnval){ [ 1120.803158] env[62814]: value = "task-4294500" [ 1120.803158] env[62814]: _type = "Task" [ 1120.803158] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.815014] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294500, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.863569] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Acquiring lock "d5df1d0e-55e6-452e-882a-053a83250b39" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1120.864467] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "d5df1d0e-55e6-452e-882a-053a83250b39" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.864467] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Acquiring lock "d5df1d0e-55e6-452e-882a-053a83250b39-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1120.864679] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "d5df1d0e-55e6-452e-882a-053a83250b39-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1120.864897] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "d5df1d0e-55e6-452e-882a-053a83250b39-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1120.869778] env[62814]: INFO nova.compute.manager [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Terminating instance [ 1120.931697] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b13b7c7-70d9-45b0-ac96-3fb7045fd6c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.940363] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd0fc57-4cfd-4c60-b35e-a43e46bad6c3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.985969] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ede6d8d6-0acd-4770-b323-c1fe0bbae081 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.997379] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b926467-39ec-4ef1-95f6-e2b4435f780b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.026369] env[62814]: DEBUG nova.compute.provider_tree [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.111720] env[62814]: DEBUG nova.compute.manager [req-b8b16c18-4991-413c-8b08-3a3171f471ed req-144a3633-fd39-4c80-aab1-3c29bc9d6868 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received event network-vif-plugged-1c6b72d7-d148-4280-981c-9bfbf768bd3b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1121.112027] env[62814]: DEBUG oslo_concurrency.lockutils [req-b8b16c18-4991-413c-8b08-3a3171f471ed req-144a3633-fd39-4c80-aab1-3c29bc9d6868 service nova] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1121.112270] env[62814]: DEBUG oslo_concurrency.lockutils [req-b8b16c18-4991-413c-8b08-3a3171f471ed req-144a3633-fd39-4c80-aab1-3c29bc9d6868 service nova] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1121.112528] env[62814]: DEBUG oslo_concurrency.lockutils [req-b8b16c18-4991-413c-8b08-3a3171f471ed req-144a3633-fd39-4c80-aab1-3c29bc9d6868 service nova] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1121.112733] env[62814]: DEBUG nova.compute.manager [req-b8b16c18-4991-413c-8b08-3a3171f471ed req-144a3633-fd39-4c80-aab1-3c29bc9d6868 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] No waiting events found dispatching network-vif-plugged-1c6b72d7-d148-4280-981c-9bfbf768bd3b {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1121.112899] env[62814]: WARNING nova.compute.manager [req-b8b16c18-4991-413c-8b08-3a3171f471ed req-144a3633-fd39-4c80-aab1-3c29bc9d6868 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received unexpected event network-vif-plugged-1c6b72d7-d148-4280-981c-9bfbf768bd3b for instance with vm_state building and task_state spawning. [ 1121.161843] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294497, 'name': CreateSnapshot_Task, 'duration_secs': 0.848651} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.162993] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1121.164667] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78927ad1-fa28-4a21-91f8-7256b46cfece {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.174218] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294499, 'name': ReconfigVM_Task, 'duration_secs': 0.21925} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.174308] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance '7e849b23-226d-4a1d-a85c-cf96964ce034' progress to 33 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1121.188058] env[62814]: DEBUG nova.network.neutron [req-575e3b8e-0450-4cb6-a37d-208805ab96ee req-c886d29a-3f68-4cb3-bf45-a0ea4ff1658b service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Updated VIF entry in instance network info cache for port 3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1121.188450] env[62814]: DEBUG nova.network.neutron [req-575e3b8e-0450-4cb6-a37d-208805ab96ee req-c886d29a-3f68-4cb3-bf45-a0ea4ff1658b service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Updating instance_info_cache with network_info: [{"id": "3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5", "address": "fa:16:3e:cb:8e:9b", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.151", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d97fdd4-3a", "ovs_interfaceid": "3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1121.314948] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294500, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.377429] env[62814]: DEBUG nova.compute.manager [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1121.377429] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-48a22d33-f560-402f-89f4-ee7897881830 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.387350] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c1e4e27-fa5e-4307-b764-0eed3ec9cd1c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.422641] env[62814]: WARNING nova.virt.vmwareapi.driver [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance d5df1d0e-55e6-452e-882a-053a83250b39 could not be found. [ 1121.422934] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1121.423383] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-174d0ea7-45be-4162-a85f-483ea3e50a34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.433978] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c7b4f72-15d9-4ebc-955e-82f67ab39bab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.469787] env[62814]: WARNING nova.virt.vmwareapi.vmops [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d5df1d0e-55e6-452e-882a-053a83250b39 could not be found. [ 1121.469787] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1121.470123] env[62814]: INFO nova.compute.manager [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Took 0.10 seconds to destroy the instance on the hypervisor. [ 1121.470779] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1121.471411] env[62814]: DEBUG nova.compute.manager [-] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1121.471411] env[62814]: DEBUG nova.network.neutron [-] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1121.480488] env[62814]: DEBUG nova.network.neutron [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Successfully updated port: 1c6b72d7-d148-4280-981c-9bfbf768bd3b {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1121.530353] env[62814]: DEBUG nova.scheduler.client.report [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1121.685343] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1121.685722] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1121.685984] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1121.686299] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1121.686596] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1121.686787] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1121.687123] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1121.687393] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1121.687668] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1121.688375] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1121.688375] env[62814]: DEBUG nova.virt.hardware [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1121.697884] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Reconfiguring VM instance instance-0000005d to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1121.698747] env[62814]: DEBUG oslo_concurrency.lockutils [req-575e3b8e-0450-4cb6-a37d-208805ab96ee req-c886d29a-3f68-4cb3-bf45-a0ea4ff1658b service nova] Releasing lock "refresh_cache-f00d414e-1dfd-43cf-9245-7cbeea8850b3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1121.711347] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1121.712045] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69261678-7f5f-4b0d-979f-48e6364481a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.736108] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5717fd48-8805-4152-9ae5-a322f4e4d497 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.749284] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1121.749284] env[62814]: value = "task-4294502" [ 1121.749284] env[62814]: _type = "Task" [ 1121.749284] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.751602] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1121.751602] env[62814]: value = "task-4294501" [ 1121.751602] env[62814]: _type = "Task" [ 1121.751602] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.771182] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294502, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.776358] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294501, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.814674] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294500, 'name': ReconfigVM_Task, 'duration_secs': 0.593704} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.814977] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Reconfigured VM instance instance-00000064 to attach disk [datastore2] e75c8039-9ff8-45f2-8770-894f2307cc14/e75c8039-9ff8-45f2-8770-894f2307cc14.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1121.815446] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95d2c276-9d4f-4380-adcb-a62e3d4493be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.822413] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for the task: (returnval){ [ 1121.822413] env[62814]: value = "task-4294503" [ 1121.822413] env[62814]: _type = "Task" [ 1121.822413] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.833159] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294503, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.989502] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1121.989628] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1121.989825] env[62814]: DEBUG nova.network.neutron [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1122.037009] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.021s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.040248] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.146s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1122.040248] env[62814]: DEBUG nova.objects.instance [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lazy-loading 'resources' on Instance uuid 1c86646f-6f38-4f8d-bea0-8a6b93680aba {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.061542] env[62814]: INFO nova.scheduler.client.report [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Deleted allocations for instance 4429ff0c-5353-4346-8bef-2c45d251f1ea [ 1122.131754] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1122.132565] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1122.264324] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294502, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.270477] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294501, 'name': ReconfigVM_Task, 'duration_secs': 0.427913} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.270859] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Reconfigured VM instance instance-0000005d to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1122.271862] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6781165-b84e-4d99-9564-6dc6d1b76e35 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.297526] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 7e849b23-226d-4a1d-a85c-cf96964ce034/7e849b23-226d-4a1d-a85c-cf96964ce034.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1122.299130] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0285615f-a404-4dc6-ba70-553adf938e37 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.314504] env[62814]: DEBUG nova.compute.manager [req-cdf328d0-1cf9-4f6c-9ddb-9afe90dd4516 req-25dfd4e2-96a2-4c2e-b427-88167ecc19d9 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received event network-changed-1c6b72d7-d148-4280-981c-9bfbf768bd3b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1122.314737] env[62814]: DEBUG nova.compute.manager [req-cdf328d0-1cf9-4f6c-9ddb-9afe90dd4516 req-25dfd4e2-96a2-4c2e-b427-88167ecc19d9 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Refreshing instance network info cache due to event network-changed-1c6b72d7-d148-4280-981c-9bfbf768bd3b. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1122.314943] env[62814]: DEBUG oslo_concurrency.lockutils [req-cdf328d0-1cf9-4f6c-9ddb-9afe90dd4516 req-25dfd4e2-96a2-4c2e-b427-88167ecc19d9 service nova] Acquiring lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.321892] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1122.321892] env[62814]: value = "task-4294504" [ 1122.321892] env[62814]: _type = "Task" [ 1122.321892] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.335795] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294504, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.341084] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294503, 'name': Rename_Task, 'duration_secs': 0.330798} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.341543] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1122.341870] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7c53050c-d1e8-494b-84ef-9c13a610e6b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.349220] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for the task: (returnval){ [ 1122.349220] env[62814]: value = "task-4294505" [ 1122.349220] env[62814]: _type = "Task" [ 1122.349220] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.358903] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294505, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.572336] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4dcff9bf-f2a2-4bc2-be73-4f9da944b163 tempest-ServerPasswordTestJSON-66340664 tempest-ServerPasswordTestJSON-66340664-project-member] Lock "4429ff0c-5353-4346-8bef-2c45d251f1ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.471s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1122.597682] env[62814]: DEBUG nova.network.neutron [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1122.637468] env[62814]: DEBUG nova.compute.manager [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1122.770743] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294502, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.832968] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294504, 'name': ReconfigVM_Task, 'duration_secs': 0.460483} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.834021] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 7e849b23-226d-4a1d-a85c-cf96964ce034/7e849b23-226d-4a1d-a85c-cf96964ce034.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1122.834500] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance '7e849b23-226d-4a1d-a85c-cf96964ce034' progress to 50 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1122.867022] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294505, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.954458] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331f65fe-92e2-46f7-978e-ca39d869ea11 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.966241] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4daf1dee-8fc1-4fd6-93c1-83795081a273 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.003399] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84375788-13eb-4d4f-9292-5f6b08f2a75d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.012651] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf30e27-2a7b-4db2-b222-a3fc2aba4465 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.033272] env[62814]: DEBUG nova.compute.provider_tree [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1123.035638] env[62814]: DEBUG nova.network.neutron [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [{"id": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "address": "fa:16:3e:e8:10:0d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c6b72d7-d1", "ovs_interfaceid": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.170944] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1123.240201] env[62814]: DEBUG nova.compute.manager [req-2489040b-0dcc-40d8-aec6-e48506178184 req-9a5aa9dc-47d1-46dc-82bd-06f22ed94377 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Received event network-vif-deleted-dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1123.240201] env[62814]: INFO nova.compute.manager [req-2489040b-0dcc-40d8-aec6-e48506178184 req-9a5aa9dc-47d1-46dc-82bd-06f22ed94377 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Neutron deleted interface dec4b0c4-68db-48fc-84e2-85becf8933bc; detaching it from the instance and deleting it from the info cache [ 1123.240201] env[62814]: DEBUG nova.network.neutron [req-2489040b-0dcc-40d8-aec6-e48506178184 req-9a5aa9dc-47d1-46dc-82bd-06f22ed94377 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.266789] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294502, 'name': CloneVM_Task, 'duration_secs': 1.482219} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.267084] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Created linked-clone VM from snapshot [ 1123.267868] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fada6dd8-4d7e-49a6-863f-b0f40f808555 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.278616] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Uploading image dc3cbb42-9340-472a-b42a-4fb9bea7421b {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1123.318669] env[62814]: DEBUG oslo_vmware.rw_handles [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1123.318669] env[62814]: value = "vm-845831" [ 1123.318669] env[62814]: _type = "VirtualMachine" [ 1123.318669] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1123.319138] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6c10a531-a259-4d12-a055-8913bb38b77c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.327173] env[62814]: DEBUG oslo_vmware.rw_handles [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lease: (returnval){ [ 1123.327173] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b376c3-5a36-b1a7-1883-bd41c583e4b3" [ 1123.327173] env[62814]: _type = "HttpNfcLease" [ 1123.327173] env[62814]: } obtained for exporting VM: (result){ [ 1123.327173] env[62814]: value = "vm-845831" [ 1123.327173] env[62814]: _type = "VirtualMachine" [ 1123.327173] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1123.327460] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the lease: (returnval){ [ 1123.327460] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b376c3-5a36-b1a7-1883-bd41c583e4b3" [ 1123.327460] env[62814]: _type = "HttpNfcLease" [ 1123.327460] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1123.334372] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1123.334372] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b376c3-5a36-b1a7-1883-bd41c583e4b3" [ 1123.334372] env[62814]: _type = "HttpNfcLease" [ 1123.334372] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1123.344881] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd68f3e-ec7d-4648-a2a8-f1f3b872f580 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.371193] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a954070-bdcf-4586-af41-48284d501961 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.380572] env[62814]: DEBUG oslo_vmware.api [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294505, 'name': PowerOnVM_Task, 'duration_secs': 0.746756} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.393491] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1123.393781] env[62814]: INFO nova.compute.manager [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Took 8.81 seconds to spawn the instance on the hypervisor. [ 1123.393974] env[62814]: DEBUG nova.compute.manager [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1123.394613] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance '7e849b23-226d-4a1d-a85c-cf96964ce034' progress to 67 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1123.398768] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda6c62f-cd60-4886-a607-d6af803e4c7c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.540035] env[62814]: DEBUG nova.scheduler.client.report [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1123.546376] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1123.547498] env[62814]: DEBUG nova.compute.manager [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Instance network_info: |[{"id": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "address": "fa:16:3e:e8:10:0d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c6b72d7-d1", "ovs_interfaceid": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1123.548357] env[62814]: DEBUG oslo_concurrency.lockutils [req-cdf328d0-1cf9-4f6c-9ddb-9afe90dd4516 req-25dfd4e2-96a2-4c2e-b427-88167ecc19d9 service nova] Acquired lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1123.548546] env[62814]: DEBUG nova.network.neutron [req-cdf328d0-1cf9-4f6c-9ddb-9afe90dd4516 req-25dfd4e2-96a2-4c2e-b427-88167ecc19d9 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Refreshing network info cache for port 1c6b72d7-d148-4280-981c-9bfbf768bd3b {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1123.549715] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:10:0d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe99da4f-5630-4afd-918b-b327193d8489', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1c6b72d7-d148-4280-981c-9bfbf768bd3b', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1123.564094] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1123.566202] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1123.566982] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eac96d42-f1ab-4292-bfca-51dd33d7ce48 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.590429] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1123.590429] env[62814]: value = "task-4294507" [ 1123.590429] env[62814]: _type = "Task" [ 1123.590429] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.600485] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294507, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.638115] env[62814]: DEBUG nova.network.neutron [-] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.744166] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-78e9cdd0-3214-492d-bbfb-6830ce397848 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.755589] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05ab1559-c05f-4743-a407-8f20370da99c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.793012] env[62814]: DEBUG nova.compute.manager [req-2489040b-0dcc-40d8-aec6-e48506178184 req-9a5aa9dc-47d1-46dc-82bd-06f22ed94377 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Detach interface failed, port_id=dec4b0c4-68db-48fc-84e2-85becf8933bc, reason: Instance d5df1d0e-55e6-452e-882a-053a83250b39 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1123.836791] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1123.836791] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b376c3-5a36-b1a7-1883-bd41c583e4b3" [ 1123.836791] env[62814]: _type = "HttpNfcLease" [ 1123.836791] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1123.837478] env[62814]: DEBUG oslo_vmware.rw_handles [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1123.837478] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b376c3-5a36-b1a7-1883-bd41c583e4b3" [ 1123.837478] env[62814]: _type = "HttpNfcLease" [ 1123.837478] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1123.841917] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1150e12d-042f-4fdb-9c83-65ec3998656e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.853741] env[62814]: DEBUG oslo_vmware.rw_handles [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a72556-da50-1819-ca87-03adf7200b1d/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1123.854127] env[62814]: DEBUG oslo_vmware.rw_handles [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a72556-da50-1819-ca87-03adf7200b1d/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1123.924601] env[62814]: INFO nova.compute.manager [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Took 34.16 seconds to build instance. [ 1123.972336] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b1069964-c974-4ab5-aed4-36f4a16d6f9e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.020603] env[62814]: DEBUG nova.network.neutron [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Port f1881a6c-5678-4b14-9eb9-f3e7f3e41810 binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1124.049474] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.010s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.051858] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.005s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1124.053929] env[62814]: INFO nova.compute.claims [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1124.091352] env[62814]: INFO nova.scheduler.client.report [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Deleted allocations for instance 1c86646f-6f38-4f8d-bea0-8a6b93680aba [ 1124.106678] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294507, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.141315] env[62814]: INFO nova.compute.manager [-] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Took 2.67 seconds to deallocate network for instance. [ 1124.340319] env[62814]: DEBUG nova.compute.manager [req-e36997fd-36ee-48e7-9149-c645e8bab442 req-2eae0a5f-792b-4116-a2a0-e883755dcfc7 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Received event network-changed-dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1124.340319] env[62814]: DEBUG nova.compute.manager [req-e36997fd-36ee-48e7-9149-c645e8bab442 req-2eae0a5f-792b-4116-a2a0-e883755dcfc7 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Refreshing instance network info cache due to event network-changed-dec4b0c4-68db-48fc-84e2-85becf8933bc. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1124.340319] env[62814]: DEBUG oslo_concurrency.lockutils [req-e36997fd-36ee-48e7-9149-c645e8bab442 req-2eae0a5f-792b-4116-a2a0-e883755dcfc7 service nova] Acquiring lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.340319] env[62814]: DEBUG oslo_concurrency.lockutils [req-e36997fd-36ee-48e7-9149-c645e8bab442 req-2eae0a5f-792b-4116-a2a0-e883755dcfc7 service nova] Acquired lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1124.340319] env[62814]: DEBUG nova.network.neutron [req-e36997fd-36ee-48e7-9149-c645e8bab442 req-2eae0a5f-792b-4116-a2a0-e883755dcfc7 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Refreshing network info cache for port dec4b0c4-68db-48fc-84e2-85becf8933bc {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1124.426449] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a5cfcf1b-4cca-4e28-86db-1cd58cc521c3 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Lock "e75c8039-9ff8-45f2-8770-894f2307cc14" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.668s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.608883] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294507, 'name': CreateVM_Task, 'duration_secs': 0.670289} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.609527] env[62814]: DEBUG oslo_concurrency.lockutils [None req-44b45664-32e9-4e07-9a95-12f3f92d2d77 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "1c86646f-6f38-4f8d-bea0-8a6b93680aba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.234s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1124.610752] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1124.613143] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.613143] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1124.613143] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1124.613143] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3998f2ba-4123-45dc-8aaa-73d3d699ff5b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.620227] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1124.620227] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52003245-da66-baaa-6262-5967069125cf" [ 1124.620227] env[62814]: _type = "Task" [ 1124.620227] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.635069] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52003245-da66-baaa-6262-5967069125cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.718236] env[62814]: INFO nova.compute.manager [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Took 0.58 seconds to detach 1 volumes for instance. [ 1124.720661] env[62814]: DEBUG nova.compute.manager [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Deleting volume: 6f2eddd0-4334-4e22-b195-b7deaa539235 {{(pid=62814) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3282}} [ 1124.735026] env[62814]: DEBUG nova.network.neutron [req-cdf328d0-1cf9-4f6c-9ddb-9afe90dd4516 req-25dfd4e2-96a2-4c2e-b427-88167ecc19d9 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updated VIF entry in instance network info cache for port 1c6b72d7-d148-4280-981c-9bfbf768bd3b. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1124.735026] env[62814]: DEBUG nova.network.neutron [req-cdf328d0-1cf9-4f6c-9ddb-9afe90dd4516 req-25dfd4e2-96a2-4c2e-b427-88167ecc19d9 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [{"id": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "address": "fa:16:3e:e8:10:0d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c6b72d7-d1", "ovs_interfaceid": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.864683] env[62814]: DEBUG nova.network.neutron [req-e36997fd-36ee-48e7-9149-c645e8bab442 req-2eae0a5f-792b-4116-a2a0-e883755dcfc7 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1125.077069] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "7e849b23-226d-4a1d-a85c-cf96964ce034-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.077069] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.077069] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.109022] env[62814]: DEBUG nova.network.neutron [req-e36997fd-36ee-48e7-9149-c645e8bab442 req-2eae0a5f-792b-4116-a2a0-e883755dcfc7 service nova] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.139221] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52003245-da66-baaa-6262-5967069125cf, 'name': SearchDatastore_Task, 'duration_secs': 0.013468} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.139554] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.139795] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1125.140083] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.140212] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1125.140403] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.140693] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6617198-10df-46c5-b77d-bd45d3496535 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.152795] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1125.153455] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1125.154812] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15188553-af54-4dd1-af10-bca977e9fa38 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.160915] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1125.160915] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52516e19-5771-759b-bcea-b3de640d60d0" [ 1125.160915] env[62814]: _type = "Task" [ 1125.160915] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.178437] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.178669] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.178950] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.179808] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1125.179808] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1125.181603] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52516e19-5771-759b-bcea-b3de640d60d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.182925] env[62814]: INFO nova.compute.manager [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Terminating instance [ 1125.237508] env[62814]: DEBUG oslo_concurrency.lockutils [req-cdf328d0-1cf9-4f6c-9ddb-9afe90dd4516 req-25dfd4e2-96a2-4c2e-b427-88167ecc19d9 service nova] Releasing lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.276244] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1125.461870] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0ba63ff-452f-425b-914b-90e9f9a32557 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.470910] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d06eab-e466-433a-a285-0394ca6e3026 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.505402] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6eeb14-64e1-436a-acb6-6f3277ca3570 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.515901] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea03aac-c928-41a3-a14b-5ed7d6d7670a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.532291] env[62814]: DEBUG nova.compute.provider_tree [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1125.611589] env[62814]: DEBUG oslo_concurrency.lockutils [req-e36997fd-36ee-48e7-9149-c645e8bab442 req-2eae0a5f-792b-4116-a2a0-e883755dcfc7 service nova] Releasing lock "refresh_cache-d5df1d0e-55e6-452e-882a-053a83250b39" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.679745] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52516e19-5771-759b-bcea-b3de640d60d0, 'name': SearchDatastore_Task, 'duration_secs': 0.018251} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.681016] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35ecce05-b753-4dea-b78d-d9dba2466041 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.689399] env[62814]: DEBUG nova.compute.manager [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1125.689977] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1125.694563] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1125.694563] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cd7d18-25a0-0211-e9ff-05ab770baaf2" [ 1125.694563] env[62814]: _type = "Task" [ 1125.694563] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.694563] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d60e234-3c88-4d34-8628-d791b17a632c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.708174] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1125.716929] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08adac53-6a5b-4154-b8b9-fb714b16454d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.720745] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cd7d18-25a0-0211-e9ff-05ab770baaf2, 'name': SearchDatastore_Task, 'duration_secs': 0.016671} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.725063] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1125.726046] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 2d7a18f3-456b-470c-a759-632e5d79a8f0/2d7a18f3-456b-470c-a759-632e5d79a8f0.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1125.728042] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f2681e4d-b452-4f40-ab7c-1dcf35d2ce6a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.732772] env[62814]: DEBUG oslo_vmware.api [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1125.732772] env[62814]: value = "task-4294509" [ 1125.732772] env[62814]: _type = "Task" [ 1125.732772] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.738749] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1125.738749] env[62814]: value = "task-4294510" [ 1125.738749] env[62814]: _type = "Task" [ 1125.738749] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.748224] env[62814]: DEBUG oslo_vmware.api [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294509, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.755915] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294510, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.038037] env[62814]: DEBUG nova.scheduler.client.report [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1126.113125] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.113125] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1126.113125] env[62814]: DEBUG nova.network.neutron [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1126.230465] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquiring lock "e75c8039-9ff8-45f2-8770-894f2307cc14" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.230465] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Lock "e75c8039-9ff8-45f2-8770-894f2307cc14" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.230465] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquiring lock "e75c8039-9ff8-45f2-8770-894f2307cc14-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1126.230465] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Lock "e75c8039-9ff8-45f2-8770-894f2307cc14-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.230465] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Lock "e75c8039-9ff8-45f2-8770-894f2307cc14-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1126.232820] env[62814]: INFO nova.compute.manager [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Terminating instance [ 1126.249360] env[62814]: DEBUG oslo_vmware.api [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294509, 'name': PowerOffVM_Task, 'duration_secs': 0.346275} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.250139] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1126.250570] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1126.257810] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e28ad3f-0547-4841-9892-7e460c55f249 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.261552] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294510, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.355791] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1126.356154] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1126.356404] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Deleting the datastore file [datastore2] cfab341e-57a1-48b4-9b6b-ceecf28c223b {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1126.357012] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6611b8fb-fb27-48a8-848a-ada0561a98d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.366252] env[62814]: DEBUG oslo_vmware.api [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for the task: (returnval){ [ 1126.366252] env[62814]: value = "task-4294512" [ 1126.366252] env[62814]: _type = "Task" [ 1126.366252] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.374957] env[62814]: DEBUG oslo_vmware.api [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294512, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.546353] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1126.546736] env[62814]: DEBUG nova.compute.manager [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1126.549657] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.993s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1126.552125] env[62814]: INFO nova.compute.claims [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1126.745945] env[62814]: DEBUG nova.compute.manager [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1126.745945] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1126.746825] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d442d34-c9a8-4960-9f51-9520597cbc3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.754067] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294510, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.701683} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.754891] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 2d7a18f3-456b-470c-a759-632e5d79a8f0/2d7a18f3-456b-470c-a759-632e5d79a8f0.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1126.755091] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1126.755343] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ad7ed688-4aff-4d3a-8e53-0db0fc967120 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.761130] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1126.765033] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31d0d201-9084-499c-8755-b1658f45ff07 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.767246] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1126.767246] env[62814]: value = "task-4294513" [ 1126.767246] env[62814]: _type = "Task" [ 1126.767246] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.772933] env[62814]: DEBUG oslo_vmware.api [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for the task: (returnval){ [ 1126.772933] env[62814]: value = "task-4294514" [ 1126.772933] env[62814]: _type = "Task" [ 1126.772933] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.781229] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294513, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.786889] env[62814]: DEBUG oslo_vmware.api [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.878514] env[62814]: DEBUG oslo_vmware.api [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Task: {'id': task-4294512, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.316438} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.878937] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1126.879505] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1126.879583] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1126.879808] env[62814]: INFO nova.compute.manager [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Took 1.19 seconds to destroy the instance on the hypervisor. [ 1126.880290] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1126.880658] env[62814]: DEBUG nova.compute.manager [-] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1126.880771] env[62814]: DEBUG nova.network.neutron [-] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1126.925850] env[62814]: DEBUG nova.network.neutron [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance_info_cache with network_info: [{"id": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "address": "fa:16:3e:35:63:2b", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1881a6c-56", "ovs_interfaceid": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.059122] env[62814]: DEBUG nova.compute.utils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1127.064167] env[62814]: DEBUG nova.compute.manager [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1127.064389] env[62814]: DEBUG nova.network.neutron [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1127.124414] env[62814]: DEBUG nova.policy [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd71116c0cfd4b509e4070ecc84613f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd82993ef7dfa4d2f8f39db3577dd321a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1127.278755] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294513, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073238} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.282922] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1127.285338] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa08599b-26ed-4b0f-bc12-393115c8e4aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.293395] env[62814]: DEBUG oslo_vmware.api [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294514, 'name': PowerOffVM_Task, 'duration_secs': 0.264732} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.303140] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1127.303449] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1127.312541] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 2d7a18f3-456b-470c-a759-632e5d79a8f0/2d7a18f3-456b-470c-a759-632e5d79a8f0.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.313598] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8945ab6-2603-4d94-ad2d-bec4498d01d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.315763] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0feb2d0-d0b9-42bf-a389-cd79a9c5e8e1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.340383] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1127.340383] env[62814]: value = "task-4294516" [ 1127.340383] env[62814]: _type = "Task" [ 1127.340383] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.350303] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294516, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.416888] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1127.417173] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1127.417409] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Deleting the datastore file [datastore2] e75c8039-9ff8-45f2-8770-894f2307cc14 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1127.421017] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90759a85-337e-4047-ad1b-5089578822b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.425176] env[62814]: DEBUG oslo_vmware.api [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for the task: (returnval){ [ 1127.425176] env[62814]: value = "task-4294517" [ 1127.425176] env[62814]: _type = "Task" [ 1127.425176] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.434853] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1127.441361] env[62814]: DEBUG oslo_vmware.api [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294517, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.442397] env[62814]: DEBUG nova.compute.manager [req-cc64ebd9-c02f-400d-884d-4b0425eb73e2 req-7a5d15b7-4c78-4f5a-ac50-63fb30474141 service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Received event network-vif-deleted-496dd017-79fb-4bcc-8fdb-32ab21f6b676 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1127.442684] env[62814]: INFO nova.compute.manager [req-cc64ebd9-c02f-400d-884d-4b0425eb73e2 req-7a5d15b7-4c78-4f5a-ac50-63fb30474141 service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Neutron deleted interface 496dd017-79fb-4bcc-8fdb-32ab21f6b676; detaching it from the instance and deleting it from the info cache [ 1127.442934] env[62814]: DEBUG nova.network.neutron [req-cc64ebd9-c02f-400d-884d-4b0425eb73e2 req-7a5d15b7-4c78-4f5a-ac50-63fb30474141 service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.451939] env[62814]: DEBUG nova.network.neutron [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Successfully created port: 78214f32-dc80-4625-b6a8-a5f4d244cfb1 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1127.564994] env[62814]: DEBUG nova.compute.manager [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1127.781752] env[62814]: DEBUG oslo_vmware.rw_handles [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52149d50-d11d-050a-d310-d42ff28d11a3/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1127.783212] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c73588-3d30-45dc-9995-9d91d168c169 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.789525] env[62814]: DEBUG nova.network.neutron [-] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.795046] env[62814]: DEBUG oslo_vmware.rw_handles [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52149d50-d11d-050a-d310-d42ff28d11a3/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1127.795418] env[62814]: ERROR oslo_vmware.rw_handles [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52149d50-d11d-050a-d310-d42ff28d11a3/disk-0.vmdk due to incomplete transfer. [ 1127.795500] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b0cffb0a-0524-49ed-8072-79977e1aa106 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.805725] env[62814]: DEBUG oslo_vmware.rw_handles [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52149d50-d11d-050a-d310-d42ff28d11a3/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1127.805939] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Uploaded image f65ac645-8a26-45ab-b54f-1c8edc8064bd to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1127.808334] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1127.808881] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-586e3f92-d7ce-454d-b18b-90d5796daf49 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.815150] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1127.815150] env[62814]: value = "task-4294518" [ 1127.815150] env[62814]: _type = "Task" [ 1127.815150] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.827136] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294518, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.851700] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294516, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.878355] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72efbf1c-8bae-405c-8653-39982eeb43f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.886233] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936d5853-85eb-4836-80e5-907facd23c85 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.919910] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d77f111-092b-49b6-b760-3d35d7dd93c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.931927] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fd6c67-8c81-4b01-a33a-299180bd1f27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.945012] env[62814]: DEBUG oslo_vmware.api [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Task: {'id': task-4294517, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.369359} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.959772] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1127.960053] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1127.960260] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1127.960510] env[62814]: INFO nova.compute.manager [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Took 1.21 seconds to destroy the instance on the hypervisor. [ 1127.960805] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1127.961269] env[62814]: DEBUG nova.compute.provider_tree [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1127.962925] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33f63d22-5ed7-4e88-b5b6-758213ccfac2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.965605] env[62814]: DEBUG nova.compute.manager [-] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1127.965717] env[62814]: DEBUG nova.network.neutron [-] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1127.976431] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d5eadf-702f-4243-bb72-40676bbaa3ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.989915] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966cc7e2-f99b-4969-a644-9a19faa8d285 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.022158] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fdaac06-666b-4943-b17f-9a32e860cacf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.025662] env[62814]: DEBUG nova.compute.manager [req-cc64ebd9-c02f-400d-884d-4b0425eb73e2 req-7a5d15b7-4c78-4f5a-ac50-63fb30474141 service nova] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Detach interface failed, port_id=496dd017-79fb-4bcc-8fdb-32ab21f6b676, reason: Instance cfab341e-57a1-48b4-9b6b-ceecf28c223b could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1128.031952] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance '7e849b23-226d-4a1d-a85c-cf96964ce034' progress to 83 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1128.295628] env[62814]: INFO nova.compute.manager [-] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Took 1.41 seconds to deallocate network for instance. [ 1128.326766] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294518, 'name': Destroy_Task, 'duration_secs': 0.398072} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.327348] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Destroyed the VM [ 1128.327592] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1128.328315] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b8fae7d9-18e1-496b-a6fe-00244d246b78 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.335551] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1128.335551] env[62814]: value = "task-4294519" [ 1128.335551] env[62814]: _type = "Task" [ 1128.335551] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.344277] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294519, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.353479] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294516, 'name': ReconfigVM_Task, 'duration_secs': 0.674071} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.353780] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 2d7a18f3-456b-470c-a759-632e5d79a8f0/2d7a18f3-456b-470c-a759-632e5d79a8f0.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.354478] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-daa020af-a5a4-4520-bbc7-39d23c5c23c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.362065] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1128.362065] env[62814]: value = "task-4294520" [ 1128.362065] env[62814]: _type = "Task" [ 1128.362065] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.370598] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294520, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.469432] env[62814]: DEBUG nova.scheduler.client.report [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1128.517624] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "d65ccde2-b31a-4032-8795-1d609fdfcc73" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1128.517907] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "d65ccde2-b31a-4032-8795-1d609fdfcc73" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1128.540148] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1128.540702] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e0a42162-8b76-4fec-aca9-36b191b37325 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.548629] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1128.548629] env[62814]: value = "task-4294521" [ 1128.548629] env[62814]: _type = "Task" [ 1128.548629] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.558636] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294521, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.578610] env[62814]: DEBUG nova.compute.manager [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1128.612852] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1128.613077] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1128.613231] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1128.613410] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1128.613554] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1128.613723] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1128.613955] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1128.614160] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1128.614336] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1128.614502] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1128.614702] env[62814]: DEBUG nova.virt.hardware [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1128.615594] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7154dd35-b557-4a25-bc2a-0aa6d128b9c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.624205] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fcef94a-9019-4b7d-ba47-7af5b817cf7e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.745264] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "29468dc9-4254-4ac6-989c-c7fa0c4cc682" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1128.745539] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "29468dc9-4254-4ac6-989c-c7fa0c4cc682" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1128.802012] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1128.802321] env[62814]: DEBUG nova.network.neutron [-] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.847593] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294519, 'name': RemoveSnapshot_Task, 'duration_secs': 0.416086} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.847945] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1128.848325] env[62814]: DEBUG nova.compute.manager [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1128.849289] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f6dbc7-8a1f-4828-a32e-c007fa77f3b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.872893] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294520, 'name': Rename_Task, 'duration_secs': 0.191092} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.874035] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1128.874035] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf843b36-5b2b-47c5-953e-37becf50a926 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.881459] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1128.881459] env[62814]: value = "task-4294522" [ 1128.881459] env[62814]: _type = "Task" [ 1128.881459] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.889675] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294522, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.974394] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1128.975115] env[62814]: DEBUG nova.compute.manager [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1128.978715] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 21.749s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1128.978949] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1128.979126] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1128.979463] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.169s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1128.987387] env[62814]: INFO nova.compute.claims [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1128.995238] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406fd35c-f936-4639-87e5-81b1bd75fec6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.016243] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3f4b04-8a44-4915-ad8e-8ee07f165c97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.024735] env[62814]: DEBUG nova.compute.manager [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1129.027604] env[62814]: DEBUG nova.compute.manager [req-6101447f-7de6-491f-8006-bab29e41086f req-903c8f80-0231-42fc-b0bf-e1c521d36709 service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Received event network-vif-plugged-78214f32-dc80-4625-b6a8-a5f4d244cfb1 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1129.027814] env[62814]: DEBUG oslo_concurrency.lockutils [req-6101447f-7de6-491f-8006-bab29e41086f req-903c8f80-0231-42fc-b0bf-e1c521d36709 service nova] Acquiring lock "eeb0255e-8c94-4564-af87-818cc8dee58e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.028028] env[62814]: DEBUG oslo_concurrency.lockutils [req-6101447f-7de6-491f-8006-bab29e41086f req-903c8f80-0231-42fc-b0bf-e1c521d36709 service nova] Lock "eeb0255e-8c94-4564-af87-818cc8dee58e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1129.028202] env[62814]: DEBUG oslo_concurrency.lockutils [req-6101447f-7de6-491f-8006-bab29e41086f req-903c8f80-0231-42fc-b0bf-e1c521d36709 service nova] Lock "eeb0255e-8c94-4564-af87-818cc8dee58e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1129.028580] env[62814]: DEBUG nova.compute.manager [req-6101447f-7de6-491f-8006-bab29e41086f req-903c8f80-0231-42fc-b0bf-e1c521d36709 service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] No waiting events found dispatching network-vif-plugged-78214f32-dc80-4625-b6a8-a5f4d244cfb1 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1129.028803] env[62814]: WARNING nova.compute.manager [req-6101447f-7de6-491f-8006-bab29e41086f req-903c8f80-0231-42fc-b0bf-e1c521d36709 service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Received unexpected event network-vif-plugged-78214f32-dc80-4625-b6a8-a5f4d244cfb1 for instance with vm_state building and task_state spawning. [ 1129.043464] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fd3e7b-5c5b-4677-a894-855cdb9cd796 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.055966] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ce3bbc-fb59-4998-aeb5-2f3b88510048 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.065976] env[62814]: DEBUG oslo_vmware.api [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294521, 'name': PowerOnVM_Task, 'duration_secs': 0.47159} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.091945] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1129.091945] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ccaf199e-fff4-4709-91fc-b73e55af42ba tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance '7e849b23-226d-4a1d-a85c-cf96964ce034' progress to 100 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1129.096028] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179284MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1129.096177] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.111792] env[62814]: DEBUG nova.network.neutron [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Successfully updated port: 78214f32-dc80-4625-b6a8-a5f4d244cfb1 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1129.248867] env[62814]: DEBUG nova.compute.manager [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1129.304450] env[62814]: INFO nova.compute.manager [-] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Took 1.34 seconds to deallocate network for instance. [ 1129.361888] env[62814]: INFO nova.compute.manager [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Shelve offloading [ 1129.391370] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294522, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.465520] env[62814]: DEBUG nova.compute.manager [req-e20e5d1a-f600-45ab-b305-8edf432b97df req-ebd62168-dd01-4b36-be98-6699cca94c0f service nova] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Received event network-vif-deleted-5c6118e7-498d-4a79-a493-8029099cd533 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1129.497300] env[62814]: DEBUG nova.compute.utils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1129.499690] env[62814]: DEBUG nova.compute.manager [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1129.499690] env[62814]: DEBUG nova.network.neutron [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1129.563106] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.577478] env[62814]: DEBUG nova.policy [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9df1642a53c940b2bfacc82e1ac975d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '446e63a428c34d38ad07d01f0c90b23c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1129.615371] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "refresh_cache-eeb0255e-8c94-4564-af87-818cc8dee58e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.615371] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "refresh_cache-eeb0255e-8c94-4564-af87-818cc8dee58e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.615371] env[62814]: DEBUG nova.network.neutron [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1129.776559] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.811904] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1129.865982] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1129.866390] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a1f5fe4-4816-441b-b582-25a65994ddb7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.875133] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1129.875133] env[62814]: value = "task-4294523" [ 1129.875133] env[62814]: _type = "Task" [ 1129.875133] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.887876] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1129.888116] env[62814]: DEBUG nova.compute.manager [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1129.889213] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd4c6d6-8538-43c7-ba5f-2b5f95e93969 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.895788] env[62814]: DEBUG oslo_concurrency.lockutils [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1129.895967] env[62814]: DEBUG oslo_concurrency.lockutils [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1129.896160] env[62814]: DEBUG nova.network.neutron [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1129.901022] env[62814]: DEBUG oslo_vmware.api [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294522, 'name': PowerOnVM_Task, 'duration_secs': 0.992914} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.901022] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1129.901022] env[62814]: INFO nova.compute.manager [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Took 9.86 seconds to spawn the instance on the hypervisor. [ 1129.901306] env[62814]: DEBUG nova.compute.manager [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1129.901971] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e53e266-ed81-4d1e-8bbc-55b2485c8010 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.003266] env[62814]: DEBUG nova.compute.manager [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1130.159299] env[62814]: DEBUG nova.network.neutron [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1130.189341] env[62814]: DEBUG nova.network.neutron [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Successfully created port: 0235e89b-ec0d-4439-817c-dd76d6987b80 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1130.292795] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce3e3edd-4e23-4836-b643-a719cfed68ce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.301132] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88754fea-654d-4371-9bcb-e95345a12c7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.335930] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7928d35-efe4-49b7-abf0-f076ca0b4324 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.344128] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db8cb6dd-9274-41dd-8643-e3c545efa383 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.358447] env[62814]: DEBUG nova.compute.provider_tree [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.380652] env[62814]: DEBUG nova.network.neutron [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Updating instance_info_cache with network_info: [{"id": "78214f32-dc80-4625-b6a8-a5f4d244cfb1", "address": "fa:16:3e:65:97:3c", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78214f32-dc", "ovs_interfaceid": "78214f32-dc80-4625-b6a8-a5f4d244cfb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.423964] env[62814]: INFO nova.compute.manager [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Took 37.64 seconds to build instance. [ 1130.726159] env[62814]: DEBUG nova.network.neutron [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Updating instance_info_cache with network_info: [{"id": "0f65290c-3126-4fa7-9407-e7ad7e8018e2", "address": "fa:16:3e:72:31:f5", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f65290c-31", "ovs_interfaceid": "0f65290c-3126-4fa7-9407-e7ad7e8018e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.864261] env[62814]: DEBUG nova.scheduler.client.report [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1130.881974] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "refresh_cache-eeb0255e-8c94-4564-af87-818cc8dee58e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1130.882648] env[62814]: DEBUG nova.compute.manager [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Instance network_info: |[{"id": "78214f32-dc80-4625-b6a8-a5f4d244cfb1", "address": "fa:16:3e:65:97:3c", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78214f32-dc", "ovs_interfaceid": "78214f32-dc80-4625-b6a8-a5f4d244cfb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1130.882968] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:97:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '78214f32-dc80-4625-b6a8-a5f4d244cfb1', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1130.891408] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1130.891408] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1130.891408] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b771bd2-f2f8-414b-b01c-39d80d238f54 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.920042] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1130.920042] env[62814]: value = "task-4294524" [ 1130.920042] env[62814]: _type = "Task" [ 1130.920042] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.930473] env[62814]: DEBUG oslo_concurrency.lockutils [None req-02873b38-fcc5-448d-8069-25ee9d16f3da tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.151s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1130.936545] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294524, 'name': CreateVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.018543] env[62814]: DEBUG nova.compute.manager [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1131.052452] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1131.052790] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1131.052989] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1131.053204] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1131.053352] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1131.054023] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1131.054023] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1131.054206] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1131.054304] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1131.054499] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1131.054691] env[62814]: DEBUG nova.virt.hardware [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1131.055837] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d327bc-c307-4e71-bdee-24112fb97421 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.065832] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f48f002-c236-49a9-95c0-fc8f94072f54 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.093166] env[62814]: DEBUG nova.compute.manager [req-bf52b02a-691b-4144-a60b-0dd78d67762e req-307a9e19-a91b-4764-9a3c-6031b974a62a service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Received event network-changed-78214f32-dc80-4625-b6a8-a5f4d244cfb1 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1131.093398] env[62814]: DEBUG nova.compute.manager [req-bf52b02a-691b-4144-a60b-0dd78d67762e req-307a9e19-a91b-4764-9a3c-6031b974a62a service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Refreshing instance network info cache due to event network-changed-78214f32-dc80-4625-b6a8-a5f4d244cfb1. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1131.093696] env[62814]: DEBUG oslo_concurrency.lockutils [req-bf52b02a-691b-4144-a60b-0dd78d67762e req-307a9e19-a91b-4764-9a3c-6031b974a62a service nova] Acquiring lock "refresh_cache-eeb0255e-8c94-4564-af87-818cc8dee58e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.093806] env[62814]: DEBUG oslo_concurrency.lockutils [req-bf52b02a-691b-4144-a60b-0dd78d67762e req-307a9e19-a91b-4764-9a3c-6031b974a62a service nova] Acquired lock "refresh_cache-eeb0255e-8c94-4564-af87-818cc8dee58e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.094022] env[62814]: DEBUG nova.network.neutron [req-bf52b02a-691b-4144-a60b-0dd78d67762e req-307a9e19-a91b-4764-9a3c-6031b974a62a service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Refreshing network info cache for port 78214f32-dc80-4625-b6a8-a5f4d244cfb1 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1131.229113] env[62814]: DEBUG oslo_concurrency.lockutils [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.376690] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.377344] env[62814]: DEBUG nova.compute.manager [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1131.380555] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.824s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.380787] env[62814]: DEBUG nova.objects.instance [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lazy-loading 'resources' on Instance uuid 08c20588-2988-4976-8631-53cf447b4877 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.431946] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "7e849b23-226d-4a1d-a85c-cf96964ce034" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1131.432183] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.432414] env[62814]: DEBUG nova.compute.manager [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Going to confirm migration 5 {{(pid=62814) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 1131.437021] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294524, 'name': CreateVM_Task, 'duration_secs': 0.444248} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.437021] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1131.437021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.437021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.437021] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1131.437021] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ad3b0ba-4ebc-40b2-a4bf-c502171d654d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.446084] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1131.446084] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5279aa40-d98c-2c78-5dbf-115239a15a15" [ 1131.446084] env[62814]: _type = "Task" [ 1131.446084] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.453019] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5279aa40-d98c-2c78-5dbf-115239a15a15, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.496655] env[62814]: DEBUG nova.compute.manager [req-953ff558-7b20-4177-812c-b21250d83e0b req-62738aa7-272a-44a6-a647-7fa05e726449 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Received event network-vif-unplugged-0f65290c-3126-4fa7-9407-e7ad7e8018e2 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1131.496969] env[62814]: DEBUG oslo_concurrency.lockutils [req-953ff558-7b20-4177-812c-b21250d83e0b req-62738aa7-272a-44a6-a647-7fa05e726449 service nova] Acquiring lock "8af98433-18eb-4a12-b18d-39b0a58821c0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1131.497212] env[62814]: DEBUG oslo_concurrency.lockutils [req-953ff558-7b20-4177-812c-b21250d83e0b req-62738aa7-272a-44a6-a647-7fa05e726449 service nova] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1131.497487] env[62814]: DEBUG oslo_concurrency.lockutils [req-953ff558-7b20-4177-812c-b21250d83e0b req-62738aa7-272a-44a6-a647-7fa05e726449 service nova] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1131.497720] env[62814]: DEBUG nova.compute.manager [req-953ff558-7b20-4177-812c-b21250d83e0b req-62738aa7-272a-44a6-a647-7fa05e726449 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] No waiting events found dispatching network-vif-unplugged-0f65290c-3126-4fa7-9407-e7ad7e8018e2 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1131.497907] env[62814]: WARNING nova.compute.manager [req-953ff558-7b20-4177-812c-b21250d83e0b req-62738aa7-272a-44a6-a647-7fa05e726449 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Received unexpected event network-vif-unplugged-0f65290c-3126-4fa7-9407-e7ad7e8018e2 for instance with vm_state shelved and task_state shelving_offloading. [ 1131.499032] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1131.500117] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb204a6-4d42-44cc-a4c9-efbc84a942ce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.508608] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1131.508864] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-274e1365-8930-4edf-a664-5536d78262e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.574771] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1131.575238] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1131.575238] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleting the datastore file [datastore2] 8af98433-18eb-4a12-b18d-39b0a58821c0 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1131.575512] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-047f56f6-b0ad-48e9-9934-1ecca5f0c562 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.582107] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1131.582107] env[62814]: value = "task-4294526" [ 1131.582107] env[62814]: _type = "Task" [ 1131.582107] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.592480] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294526, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.881550] env[62814]: DEBUG nova.network.neutron [req-bf52b02a-691b-4144-a60b-0dd78d67762e req-307a9e19-a91b-4764-9a3c-6031b974a62a service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Updated VIF entry in instance network info cache for port 78214f32-dc80-4625-b6a8-a5f4d244cfb1. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1131.881896] env[62814]: DEBUG nova.network.neutron [req-bf52b02a-691b-4144-a60b-0dd78d67762e req-307a9e19-a91b-4764-9a3c-6031b974a62a service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Updating instance_info_cache with network_info: [{"id": "78214f32-dc80-4625-b6a8-a5f4d244cfb1", "address": "fa:16:3e:65:97:3c", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap78214f32-dc", "ovs_interfaceid": "78214f32-dc80-4625-b6a8-a5f4d244cfb1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.884299] env[62814]: DEBUG nova.compute.utils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1131.892387] env[62814]: DEBUG nova.compute.manager [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1131.892387] env[62814]: DEBUG nova.network.neutron [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1131.952794] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5279aa40-d98c-2c78-5dbf-115239a15a15, 'name': SearchDatastore_Task, 'duration_secs': 0.018701} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.953135] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1131.953380] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1131.953619] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.953779] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1131.954443] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1131.956893] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-89ae8a95-a9be-4610-a390-5536c30effed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.961334] env[62814]: DEBUG nova.policy [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8497da37b9de4640826067a536d8a358', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2b3ba1c9c7624abfb3e0bd880dc40166', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1131.970330] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1131.970514] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1131.971398] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e67979a-1e78-4865-85f3-1d9871950080 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.980136] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1131.980136] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529539a5-714f-8d60-c3ed-d62d1ee75c7a" [ 1131.980136] env[62814]: _type = "Task" [ 1131.980136] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.990018] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529539a5-714f-8d60-c3ed-d62d1ee75c7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.001057] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.001057] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1132.001701] env[62814]: DEBUG nova.network.neutron [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1132.001701] env[62814]: DEBUG nova.objects.instance [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lazy-loading 'info_cache' on Instance uuid 7e849b23-226d-4a1d-a85c-cf96964ce034 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1132.091006] env[62814]: DEBUG nova.network.neutron [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Successfully updated port: 0235e89b-ec0d-4439-817c-dd76d6987b80 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1132.107984] env[62814]: DEBUG oslo_vmware.api [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294526, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.418234} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.112805] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1132.113943] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1132.113943] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1132.146339] env[62814]: INFO nova.scheduler.client.report [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted allocations for instance 8af98433-18eb-4a12-b18d-39b0a58821c0 [ 1132.239389] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5778ae2-2713-4bea-b286-087bb71febbd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.248474] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6220999-0e61-4da8-bc59-0565d0de537a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.284329] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bec35e4-575d-4ba2-82fc-11d833df11cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.292026] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9be029-c5ec-4950-9c38-4cee818e9f23 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.306345] env[62814]: DEBUG nova.compute.provider_tree [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.318476] env[62814]: DEBUG nova.network.neutron [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Successfully created port: 446b3cd1-e546-440b-bf99-c5bfb3dd4878 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1132.388894] env[62814]: DEBUG oslo_concurrency.lockutils [req-bf52b02a-691b-4144-a60b-0dd78d67762e req-307a9e19-a91b-4764-9a3c-6031b974a62a service nova] Releasing lock "refresh_cache-eeb0255e-8c94-4564-af87-818cc8dee58e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1132.389545] env[62814]: DEBUG nova.compute.manager [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1132.491208] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529539a5-714f-8d60-c3ed-d62d1ee75c7a, 'name': SearchDatastore_Task, 'duration_secs': 0.018231} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.492025] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-092887e3-abdb-42dc-ba2c-66da0024a720 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.497616] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1132.497616] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5229d582-3dab-c989-bad4-a3cee3d39192" [ 1132.497616] env[62814]: _type = "Task" [ 1132.497616] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.507463] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5229d582-3dab-c989-bad4-a3cee3d39192, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.597783] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.597992] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1132.598216] env[62814]: DEBUG nova.network.neutron [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1132.651603] env[62814]: DEBUG oslo_concurrency.lockutils [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1132.809985] env[62814]: DEBUG nova.scheduler.client.report [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1133.010666] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5229d582-3dab-c989-bad4-a3cee3d39192, 'name': SearchDatastore_Task, 'duration_secs': 0.016835} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.011074] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1133.011314] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] eeb0255e-8c94-4564-af87-818cc8dee58e/eeb0255e-8c94-4564-af87-818cc8dee58e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1133.011914] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-455c948f-3cd2-4f51-9024-384c598673f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.019238] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1133.019238] env[62814]: value = "task-4294527" [ 1133.019238] env[62814]: _type = "Task" [ 1133.019238] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.028387] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294527, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.173319] env[62814]: DEBUG nova.network.neutron [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1133.314756] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.934s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.316937] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.148s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1133.324207] env[62814]: DEBUG nova.objects.instance [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lazy-loading 'resources' on Instance uuid e727eeb1-c5d0-4591-80bb-31746bf976a2 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.347048] env[62814]: INFO nova.scheduler.client.report [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Deleted allocations for instance 08c20588-2988-4976-8631-53cf447b4877 [ 1133.399700] env[62814]: DEBUG nova.compute.manager [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1133.427480] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1133.427721] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1133.428946] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1133.428946] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1133.428946] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1133.428946] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1133.428946] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1133.428946] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1133.429507] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1133.429507] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1133.429507] env[62814]: DEBUG nova.virt.hardware [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1133.430637] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cdeee00-008f-48d4-9a34-cec8602386df {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.436310] env[62814]: DEBUG nova.compute.manager [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received event network-changed-1c6b72d7-d148-4280-981c-9bfbf768bd3b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1133.436568] env[62814]: DEBUG nova.compute.manager [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Refreshing instance network info cache due to event network-changed-1c6b72d7-d148-4280-981c-9bfbf768bd3b. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1133.436711] env[62814]: DEBUG oslo_concurrency.lockutils [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] Acquiring lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.436859] env[62814]: DEBUG oslo_concurrency.lockutils [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] Acquired lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1133.437030] env[62814]: DEBUG nova.network.neutron [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Refreshing network info cache for port 1c6b72d7-d148-4280-981c-9bfbf768bd3b {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1133.439586] env[62814]: DEBUG nova.network.neutron [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance_info_cache with network_info: [{"id": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "address": "fa:16:3e:35:63:2b", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf1881a6c-56", "ovs_interfaceid": "f1881a6c-5678-4b14-9eb9-f3e7f3e41810", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.446751] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0089a5-26f2-41f7-b9b8-290ce26529eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.455124] env[62814]: DEBUG nova.network.neutron [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance_info_cache with network_info: [{"id": "0235e89b-ec0d-4439-817c-dd76d6987b80", "address": "fa:16:3e:46:9d:6d", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0235e89b-ec", "ovs_interfaceid": "0235e89b-ec0d-4439-817c-dd76d6987b80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1133.536864] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294527, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.540740] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "8af98433-18eb-4a12-b18d-39b0a58821c0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1133.681347] env[62814]: DEBUG nova.compute.manager [req-2baf0c27-a1ba-4a97-817d-31af389adcbe req-cccdd9a2-3cff-4c8a-bc60-e7d957ff7696 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Received event network-changed-0f65290c-3126-4fa7-9407-e7ad7e8018e2 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1133.681542] env[62814]: DEBUG nova.compute.manager [req-2baf0c27-a1ba-4a97-817d-31af389adcbe req-cccdd9a2-3cff-4c8a-bc60-e7d957ff7696 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Refreshing instance network info cache due to event network-changed-0f65290c-3126-4fa7-9407-e7ad7e8018e2. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1133.681755] env[62814]: DEBUG oslo_concurrency.lockutils [req-2baf0c27-a1ba-4a97-817d-31af389adcbe req-cccdd9a2-3cff-4c8a-bc60-e7d957ff7696 service nova] Acquiring lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.681900] env[62814]: DEBUG oslo_concurrency.lockutils [req-2baf0c27-a1ba-4a97-817d-31af389adcbe req-cccdd9a2-3cff-4c8a-bc60-e7d957ff7696 service nova] Acquired lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1133.682322] env[62814]: DEBUG nova.network.neutron [req-2baf0c27-a1ba-4a97-817d-31af389adcbe req-cccdd9a2-3cff-4c8a-bc60-e7d957ff7696 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Refreshing network info cache for port 0f65290c-3126-4fa7-9407-e7ad7e8018e2 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1133.865262] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a7bc17fe-38fc-4601-82da-5e887df46546 tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "08c20588-2988-4976-8631-53cf447b4877" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.764s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1133.945761] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-7e849b23-226d-4a1d-a85c-cf96964ce034" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1133.947023] env[62814]: DEBUG nova.objects.instance [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lazy-loading 'migration_context' on Instance uuid 7e849b23-226d-4a1d-a85c-cf96964ce034 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.967838] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1133.968325] env[62814]: DEBUG nova.compute.manager [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Instance network_info: |[{"id": "0235e89b-ec0d-4439-817c-dd76d6987b80", "address": "fa:16:3e:46:9d:6d", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0235e89b-ec", "ovs_interfaceid": "0235e89b-ec0d-4439-817c-dd76d6987b80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1133.971231] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:46:9d:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b83383f-ed7a-4efd-aef7-aa8c15649d07', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0235e89b-ec0d-4439-817c-dd76d6987b80', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1133.979044] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1133.979813] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1133.985422] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6fdcf73-f036-4ca5-b6ff-5a6c521ebecc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.013752] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1134.013752] env[62814]: value = "task-4294528" [ 1134.013752] env[62814]: _type = "Task" [ 1134.013752] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.026176] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294528, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.027022] env[62814]: DEBUG nova.network.neutron [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Successfully updated port: 446b3cd1-e546-440b-bf99-c5bfb3dd4878 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1134.034697] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294527, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.80722} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.034965] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] eeb0255e-8c94-4564-af87-818cc8dee58e/eeb0255e-8c94-4564-af87-818cc8dee58e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1134.035198] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1134.035455] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a804474e-024b-451a-ad20-334b29d90d01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.044304] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1134.044304] env[62814]: value = "task-4294529" [ 1134.044304] env[62814]: _type = "Task" [ 1134.044304] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.060674] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.158112] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649f25ad-f73a-49b1-828e-3b1fcbe32457 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.166698] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c7b51d-e1ed-4c62-9f6d-2984f32a967e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.203586] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fdcbff1-6763-4d31-9f6e-3568e9cf819b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.214261] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-174d3889-4375-4be8-b8d3-494120ec1905 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.228234] env[62814]: DEBUG nova.compute.provider_tree [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.312062] env[62814]: DEBUG nova.network.neutron [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updated VIF entry in instance network info cache for port 1c6b72d7-d148-4280-981c-9bfbf768bd3b. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1134.312454] env[62814]: DEBUG nova.network.neutron [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [{"id": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "address": "fa:16:3e:e8:10:0d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c6b72d7-d1", "ovs_interfaceid": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.428763] env[62814]: DEBUG nova.network.neutron [req-2baf0c27-a1ba-4a97-817d-31af389adcbe req-cccdd9a2-3cff-4c8a-bc60-e7d957ff7696 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Updated VIF entry in instance network info cache for port 0f65290c-3126-4fa7-9407-e7ad7e8018e2. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1134.429176] env[62814]: DEBUG nova.network.neutron [req-2baf0c27-a1ba-4a97-817d-31af389adcbe req-cccdd9a2-3cff-4c8a-bc60-e7d957ff7696 service nova] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Updating instance_info_cache with network_info: [{"id": "0f65290c-3126-4fa7-9407-e7ad7e8018e2", "address": "fa:16:3e:72:31:f5", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": null, "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0f65290c-31", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.449804] env[62814]: DEBUG nova.objects.base [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Object Instance<7e849b23-226d-4a1d-a85c-cf96964ce034> lazy-loaded attributes: info_cache,migration_context {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1134.451106] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c82ee1-1eee-4541-83d3-4ae3709fdd2d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.473175] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13f55acc-1d69-4192-a9db-364aef56c336 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.479050] env[62814]: DEBUG oslo_vmware.api [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1134.479050] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f1803-57d8-a714-c28b-4541dca11118" [ 1134.479050] env[62814]: _type = "Task" [ 1134.479050] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.487114] env[62814]: DEBUG oslo_vmware.api [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f1803-57d8-a714-c28b-4541dca11118, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.524435] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294528, 'name': CreateVM_Task, 'duration_secs': 0.395695} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.524616] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1134.525336] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.525508] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1134.525832] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1134.526108] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f4778e8-f12e-4bbd-96d9-413f973a10b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.530207] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "refresh_cache-33ab2d9e-e960-40ae-acf9-4fea0d9c830e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.531291] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquired lock "refresh_cache-33ab2d9e-e960-40ae-acf9-4fea0d9c830e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1134.531291] env[62814]: DEBUG nova.network.neutron [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1134.533667] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1134.533667] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a31241-735a-2c1b-ef6c-307be0b3561c" [ 1134.533667] env[62814]: _type = "Task" [ 1134.533667] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.545313] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a31241-735a-2c1b-ef6c-307be0b3561c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.554840] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084613} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.555372] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1134.556098] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50b28cee-929f-4e64-939c-a5e29df37ccb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.580720] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] eeb0255e-8c94-4564-af87-818cc8dee58e/eeb0255e-8c94-4564-af87-818cc8dee58e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.581080] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2b2f459-debe-4c7d-acc6-5d8746e471be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.604745] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1134.604745] env[62814]: value = "task-4294530" [ 1134.604745] env[62814]: _type = "Task" [ 1134.604745] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.613291] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294530, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.630431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.630686] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.630926] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.631133] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.631397] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1134.633877] env[62814]: INFO nova.compute.manager [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Terminating instance [ 1134.731749] env[62814]: DEBUG nova.scheduler.client.report [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1134.815216] env[62814]: DEBUG oslo_concurrency.lockutils [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] Releasing lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1134.815500] env[62814]: DEBUG nova.compute.manager [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Received event network-vif-plugged-0235e89b-ec0d-4439-817c-dd76d6987b80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1134.815693] env[62814]: DEBUG oslo_concurrency.lockutils [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] Acquiring lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1134.815894] env[62814]: DEBUG oslo_concurrency.lockutils [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1134.816073] env[62814]: DEBUG oslo_concurrency.lockutils [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1134.816245] env[62814]: DEBUG nova.compute.manager [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] No waiting events found dispatching network-vif-plugged-0235e89b-ec0d-4439-817c-dd76d6987b80 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1134.816412] env[62814]: WARNING nova.compute.manager [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Received unexpected event network-vif-plugged-0235e89b-ec0d-4439-817c-dd76d6987b80 for instance with vm_state building and task_state spawning. [ 1134.816575] env[62814]: DEBUG nova.compute.manager [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Received event network-changed-0235e89b-ec0d-4439-817c-dd76d6987b80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1134.816731] env[62814]: DEBUG nova.compute.manager [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Refreshing instance network info cache due to event network-changed-0235e89b-ec0d-4439-817c-dd76d6987b80. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1134.816921] env[62814]: DEBUG oslo_concurrency.lockutils [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] Acquiring lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1134.817071] env[62814]: DEBUG oslo_concurrency.lockutils [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] Acquired lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1134.817231] env[62814]: DEBUG nova.network.neutron [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Refreshing network info cache for port 0235e89b-ec0d-4439-817c-dd76d6987b80 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1134.859678] env[62814]: DEBUG oslo_vmware.rw_handles [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a72556-da50-1819-ca87-03adf7200b1d/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1134.860863] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c52f6df-d608-4d68-83f4-917686aea339 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.866723] env[62814]: DEBUG oslo_vmware.rw_handles [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a72556-da50-1819-ca87-03adf7200b1d/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1134.866891] env[62814]: ERROR oslo_vmware.rw_handles [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a72556-da50-1819-ca87-03adf7200b1d/disk-0.vmdk due to incomplete transfer. [ 1134.867108] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c9e04b7e-7491-4c6d-9aee-b406729ad391 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.875769] env[62814]: DEBUG oslo_vmware.rw_handles [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52a72556-da50-1819-ca87-03adf7200b1d/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1134.875957] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Uploaded image dc3cbb42-9340-472a-b42a-4fb9bea7421b to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1134.877741] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1134.877963] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b954056f-44d6-4b28-b8e7-c91eb29275c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.884107] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1134.884107] env[62814]: value = "task-4294531" [ 1134.884107] env[62814]: _type = "Task" [ 1134.884107] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.894449] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294531, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.932577] env[62814]: DEBUG oslo_concurrency.lockutils [req-2baf0c27-a1ba-4a97-817d-31af389adcbe req-cccdd9a2-3cff-4c8a-bc60-e7d957ff7696 service nova] Releasing lock "refresh_cache-8af98433-18eb-4a12-b18d-39b0a58821c0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1134.990025] env[62814]: DEBUG oslo_vmware.api [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523f1803-57d8-a714-c28b-4541dca11118, 'name': SearchDatastore_Task, 'duration_secs': 0.017452} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.990025] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.046257] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a31241-735a-2c1b-ef6c-307be0b3561c, 'name': SearchDatastore_Task, 'duration_secs': 0.013597} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.046680] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1135.046928] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1135.048042] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.048042] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.048042] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1135.048042] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-072e1ae2-b471-40a4-a37f-81fd4dc77670 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.057226] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1135.057493] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1135.059093] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-abaf2328-b1ee-453a-b477-d015c03e73a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.064456] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1135.064456] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52428aaf-64e8-4ad1-7454-723a38221744" [ 1135.064456] env[62814]: _type = "Task" [ 1135.064456] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.073797] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52428aaf-64e8-4ad1-7454-723a38221744, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.082949] env[62814]: DEBUG nova.network.neutron [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1135.113757] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294530, 'name': ReconfigVM_Task, 'duration_secs': 0.290215} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.114043] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Reconfigured VM instance instance-00000066 to attach disk [datastore2] eeb0255e-8c94-4564-af87-818cc8dee58e/eeb0255e-8c94-4564-af87-818cc8dee58e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1135.114668] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a851e7ce-6388-4080-9a0f-2dbd68cf0aef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.122039] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1135.122039] env[62814]: value = "task-4294532" [ 1135.122039] env[62814]: _type = "Task" [ 1135.122039] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.129941] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294532, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.137592] env[62814]: DEBUG nova.compute.manager [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1135.137788] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1135.138527] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7f2757a-6b89-4422-8ae6-180a8e2818ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.144768] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1135.147251] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08ae76af-b50a-4706-b74e-30ae35c43429 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.153158] env[62814]: DEBUG oslo_vmware.api [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1135.153158] env[62814]: value = "task-4294533" [ 1135.153158] env[62814]: _type = "Task" [ 1135.153158] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.160511] env[62814]: DEBUG oslo_vmware.api [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.209613] env[62814]: DEBUG nova.network.neutron [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Updating instance_info_cache with network_info: [{"id": "446b3cd1-e546-440b-bf99-c5bfb3dd4878", "address": "fa:16:3e:d1:71:e0", "network": {"id": "5363651c-6ade-4a4c-b762-9fc7cbbcf5ea", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1833251571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b3ba1c9c7624abfb3e0bd880dc40166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap446b3cd1-e5", "ovs_interfaceid": "446b3cd1-e546-440b-bf99-c5bfb3dd4878", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.237528] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.920s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.239884] env[62814]: DEBUG oslo_concurrency.lockutils [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.718s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.263573] env[62814]: INFO nova.scheduler.client.report [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Deleted allocations for instance e727eeb1-c5d0-4591-80bb-31746bf976a2 [ 1135.396373] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294531, 'name': Destroy_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.478780] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b66ef5-c166-4530-ba92-580d35ad50eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.488263] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20704b4f-f612-4b18-84c6-1c16b903db48 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.520461] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81b8567-e3ed-44df-9eef-43b30b52ad61 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.528115] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25a5caa-e166-4b92-970c-3073f003f4a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.546948] env[62814]: DEBUG nova.compute.provider_tree [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1135.574737] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52428aaf-64e8-4ad1-7454-723a38221744, 'name': SearchDatastore_Task, 'duration_secs': 0.044664} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.575595] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-98054096-7502-4cd2-85bf-6d18f7bfab36 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.581295] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1135.581295] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522b9c4b-268a-782c-908c-84f112b98e37" [ 1135.581295] env[62814]: _type = "Task" [ 1135.581295] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.588724] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522b9c4b-268a-782c-908c-84f112b98e37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.617888] env[62814]: DEBUG nova.network.neutron [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updated VIF entry in instance network info cache for port 0235e89b-ec0d-4439-817c-dd76d6987b80. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1135.618056] env[62814]: DEBUG nova.network.neutron [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance_info_cache with network_info: [{"id": "0235e89b-ec0d-4439-817c-dd76d6987b80", "address": "fa:16:3e:46:9d:6d", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0235e89b-ec", "ovs_interfaceid": "0235e89b-ec0d-4439-817c-dd76d6987b80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.632070] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294532, 'name': Rename_Task, 'duration_secs': 0.179863} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.633098] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.633464] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ae3a7aa-7571-4f2e-b9b1-c2caa4218f68 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.642448] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1135.642448] env[62814]: value = "task-4294534" [ 1135.642448] env[62814]: _type = "Task" [ 1135.642448] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.650747] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294534, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.663188] env[62814]: DEBUG oslo_vmware.api [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294533, 'name': PowerOffVM_Task, 'duration_secs': 0.199945} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.663188] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1135.663188] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1135.663188] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6234181c-e75b-4a15-9c31-eaacb73e73f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.712209] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Releasing lock "refresh_cache-33ab2d9e-e960-40ae-acf9-4fea0d9c830e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1135.712564] env[62814]: DEBUG nova.compute.manager [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Instance network_info: |[{"id": "446b3cd1-e546-440b-bf99-c5bfb3dd4878", "address": "fa:16:3e:d1:71:e0", "network": {"id": "5363651c-6ade-4a4c-b762-9fc7cbbcf5ea", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1833251571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b3ba1c9c7624abfb3e0bd880dc40166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap446b3cd1-e5", "ovs_interfaceid": "446b3cd1-e546-440b-bf99-c5bfb3dd4878", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1135.713033] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:71:e0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '50cf0a70-948d-4611-af05-94c1483064ed', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '446b3cd1-e546-440b-bf99-c5bfb3dd4878', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1135.720840] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1135.721086] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1135.721322] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-51d07ba2-ecfe-47af-9946-cf3d7ace5951 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.738452] env[62814]: DEBUG nova.compute.manager [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Received event network-vif-plugged-446b3cd1-e546-440b-bf99-c5bfb3dd4878 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1135.738669] env[62814]: DEBUG oslo_concurrency.lockutils [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1135.738877] env[62814]: DEBUG oslo_concurrency.lockutils [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1135.739064] env[62814]: DEBUG oslo_concurrency.lockutils [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.739242] env[62814]: DEBUG nova.compute.manager [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] No waiting events found dispatching network-vif-plugged-446b3cd1-e546-440b-bf99-c5bfb3dd4878 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1135.739413] env[62814]: WARNING nova.compute.manager [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Received unexpected event network-vif-plugged-446b3cd1-e546-440b-bf99-c5bfb3dd4878 for instance with vm_state building and task_state spawning. [ 1135.739570] env[62814]: DEBUG nova.compute.manager [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Received event network-changed-446b3cd1-e546-440b-bf99-c5bfb3dd4878 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1135.739721] env[62814]: DEBUG nova.compute.manager [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Refreshing instance network info cache due to event network-changed-446b3cd1-e546-440b-bf99-c5bfb3dd4878. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1135.739934] env[62814]: DEBUG oslo_concurrency.lockutils [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] Acquiring lock "refresh_cache-33ab2d9e-e960-40ae-acf9-4fea0d9c830e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.740050] env[62814]: DEBUG oslo_concurrency.lockutils [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] Acquired lock "refresh_cache-33ab2d9e-e960-40ae-acf9-4fea0d9c830e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1135.740211] env[62814]: DEBUG nova.network.neutron [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Refreshing network info cache for port 446b3cd1-e546-440b-bf99-c5bfb3dd4878 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1135.749602] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1135.749602] env[62814]: value = "task-4294536" [ 1135.749602] env[62814]: _type = "Task" [ 1135.749602] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.759955] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294536, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.762144] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1135.762360] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1135.762538] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Deleting the datastore file [datastore2] 0c6c598f-ef5a-4e91-b811-cd3d8f072647 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1135.762835] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0974e3e3-c4a8-4c66-97fd-74e05ded95f3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.771740] env[62814]: DEBUG oslo_vmware.api [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for the task: (returnval){ [ 1135.771740] env[62814]: value = "task-4294537" [ 1135.771740] env[62814]: _type = "Task" [ 1135.771740] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.772242] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ba98a21-d381-459f-8d88-4bced78e6a53 tempest-ServerRescueTestJSON-199202121 tempest-ServerRescueTestJSON-199202121-project-member] Lock "e727eeb1-c5d0-4591-80bb-31746bf976a2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.108s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1135.784777] env[62814]: DEBUG oslo_vmware.api [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294537, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.896150] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294531, 'name': Destroy_Task, 'duration_secs': 0.739599} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.896418] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Destroyed the VM [ 1135.896661] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1135.896912] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3fe4fcc6-89ff-4b39-9fbd-92577d7311f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.904246] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1135.904246] env[62814]: value = "task-4294538" [ 1135.904246] env[62814]: _type = "Task" [ 1135.904246] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.915352] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294538, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.051547] env[62814]: DEBUG nova.scheduler.client.report [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1136.098474] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522b9c4b-268a-782c-908c-84f112b98e37, 'name': SearchDatastore_Task, 'duration_secs': 0.010887} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.098867] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.099219] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b1d88997-e52f-41bd-b1b4-dd096d20d60a/b1d88997-e52f-41bd-b1b4-dd096d20d60a.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1136.099513] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e6ef5e0-b79f-49e1-8590-90c83d19a4aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.107564] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1136.107564] env[62814]: value = "task-4294539" [ 1136.107564] env[62814]: _type = "Task" [ 1136.107564] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.117148] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294539, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.121114] env[62814]: DEBUG oslo_concurrency.lockutils [req-a01fb8ca-6180-46f2-ab8b-c17ad18e259a req-1e6cdd1d-0ef9-4b81-9c17-0da004cef77a service nova] Releasing lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.152954] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294534, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.261230] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294536, 'name': CreateVM_Task, 'duration_secs': 0.421833} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.261444] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1136.262215] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.262436] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.262860] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1136.263462] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8465011d-e9b4-4cd3-b295-34c4d4ddcf3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.268367] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1136.268367] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52155c09-53a0-8259-b974-bd76bde4811c" [ 1136.268367] env[62814]: _type = "Task" [ 1136.268367] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.279250] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52155c09-53a0-8259-b974-bd76bde4811c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.284888] env[62814]: DEBUG oslo_vmware.api [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Task: {'id': task-4294537, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228566} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.285163] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1136.285343] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1136.285521] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1136.285701] env[62814]: INFO nova.compute.manager [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1136.285946] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1136.286156] env[62814]: DEBUG nova.compute.manager [-] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1136.286243] env[62814]: DEBUG nova.network.neutron [-] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1136.415015] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294538, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.559572] env[62814]: DEBUG oslo_concurrency.lockutils [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.319s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1136.559854] env[62814]: INFO nova.compute.manager [None req-443e90c3-2b21-4a55-82e1-ff9eef81d61a tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Successfully reverted task state from rebuilding on failure for instance. [ 1136.567869] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.397s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1136.569704] env[62814]: INFO nova.compute.claims [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1136.622920] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294539, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.655595] env[62814]: DEBUG oslo_vmware.api [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294534, 'name': PowerOnVM_Task, 'duration_secs': 0.517573} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.655595] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.655595] env[62814]: INFO nova.compute.manager [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Took 8.07 seconds to spawn the instance on the hypervisor. [ 1136.655595] env[62814]: DEBUG nova.compute.manager [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1136.655595] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c51e1f40-bc28-420c-92a1-260e8ca3c714 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.702587] env[62814]: DEBUG nova.network.neutron [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Updated VIF entry in instance network info cache for port 446b3cd1-e546-440b-bf99-c5bfb3dd4878. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1136.703042] env[62814]: DEBUG nova.network.neutron [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Updating instance_info_cache with network_info: [{"id": "446b3cd1-e546-440b-bf99-c5bfb3dd4878", "address": "fa:16:3e:d1:71:e0", "network": {"id": "5363651c-6ade-4a4c-b762-9fc7cbbcf5ea", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1833251571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b3ba1c9c7624abfb3e0bd880dc40166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap446b3cd1-e5", "ovs_interfaceid": "446b3cd1-e546-440b-bf99-c5bfb3dd4878", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.783462] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52155c09-53a0-8259-b974-bd76bde4811c, 'name': SearchDatastore_Task, 'duration_secs': 0.012079} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.783942] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1136.784323] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1136.784681] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.784916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1136.785227] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1136.785977] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32d681ac-1307-47e0-b8a1-31595ff0caf3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.797888] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1136.798191] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1136.799342] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3480ab75-aaa3-424b-b65e-5e9ad70354d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.806416] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1136.806416] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528d0a58-9d88-ef3a-0120-5d5eed7000c9" [ 1136.806416] env[62814]: _type = "Task" [ 1136.806416] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.817834] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528d0a58-9d88-ef3a-0120-5d5eed7000c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.858708] env[62814]: DEBUG nova.compute.manager [req-5c5dede2-6fbb-4513-a207-d6b65466f8dc req-90fab1ac-5ac8-4481-92b1-d7d0ae4d4e14 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Received event network-vif-deleted-b3b985f6-c18c-47da-a863-93d8e8a87f19 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1136.859330] env[62814]: INFO nova.compute.manager [req-5c5dede2-6fbb-4513-a207-d6b65466f8dc req-90fab1ac-5ac8-4481-92b1-d7d0ae4d4e14 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Neutron deleted interface b3b985f6-c18c-47da-a863-93d8e8a87f19; detaching it from the instance and deleting it from the info cache [ 1136.859660] env[62814]: DEBUG nova.network.neutron [req-5c5dede2-6fbb-4513-a207-d6b65466f8dc req-90fab1ac-5ac8-4481-92b1-d7d0ae4d4e14 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.922489] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294538, 'name': RemoveSnapshot_Task, 'duration_secs': 0.843115} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.922998] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1136.923282] env[62814]: DEBUG nova.compute.manager [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1136.924767] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0becac-e9a6-4a3e-a491-b79cb3b9590f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.119437] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294539, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56204} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.120553] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b1d88997-e52f-41bd-b1b4-dd096d20d60a/b1d88997-e52f-41bd-b1b4-dd096d20d60a.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1137.122155] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1137.122155] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-379ab7ca-b174-4428-ab46-7a3c963ec7f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.129740] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1137.129740] env[62814]: value = "task-4294540" [ 1137.129740] env[62814]: _type = "Task" [ 1137.129740] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.138992] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294540, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.177496] env[62814]: INFO nova.compute.manager [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Took 33.15 seconds to build instance. [ 1137.187251] env[62814]: DEBUG nova.network.neutron [-] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.205131] env[62814]: DEBUG oslo_concurrency.lockutils [req-691131b7-1c36-4784-ac1f-8cbe28d57e18 req-0fa138d4-fa90-450e-bc2e-52acb0ef999d service nova] Releasing lock "refresh_cache-33ab2d9e-e960-40ae-acf9-4fea0d9c830e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.321024] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528d0a58-9d88-ef3a-0120-5d5eed7000c9, 'name': SearchDatastore_Task, 'duration_secs': 0.037846} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.321024] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eee2e0ee-fc8a-4f46-89e8-ec44e43f903b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.327515] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1137.327515] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d1e97a-391e-0842-ac9b-3d85bc9b1b7e" [ 1137.327515] env[62814]: _type = "Task" [ 1137.327515] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.336421] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d1e97a-391e-0842-ac9b-3d85bc9b1b7e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.363489] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-883fa371-df91-4085-92f0-26f6bdc7601b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.374269] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad2e9b8-adf2-403e-8ec2-744a71fbdb4e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.409424] env[62814]: DEBUG nova.compute.manager [req-5c5dede2-6fbb-4513-a207-d6b65466f8dc req-90fab1ac-5ac8-4481-92b1-d7d0ae4d4e14 service nova] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Detach interface failed, port_id=b3b985f6-c18c-47da-a863-93d8e8a87f19, reason: Instance 0c6c598f-ef5a-4e91-b811-cd3d8f072647 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1137.447679] env[62814]: INFO nova.compute.manager [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Shelve offloading [ 1137.644949] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294540, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06759} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.644949] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1137.644949] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386d4352-1662-4086-91e7-12d5a7b5e780 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.667636] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] b1d88997-e52f-41bd-b1b4-dd096d20d60a/b1d88997-e52f-41bd-b1b4-dd096d20d60a.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1137.670496] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-982c86a1-f3de-4e1a-a450-6c97bb87ec09 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.686035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "eeb0255e-8c94-4564-af87-818cc8dee58e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1137.686200] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1565b678-4632-4cbf-b9c6-4682f54e2be0 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "eeb0255e-8c94-4564-af87-818cc8dee58e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.665s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1137.686323] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "eeb0255e-8c94-4564-af87-818cc8dee58e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1137.686524] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "eeb0255e-8c94-4564-af87-818cc8dee58e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1137.686714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "eeb0255e-8c94-4564-af87-818cc8dee58e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1137.686877] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "eeb0255e-8c94-4564-af87-818cc8dee58e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1137.688962] env[62814]: INFO nova.compute.manager [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Terminating instance [ 1137.690383] env[62814]: INFO nova.compute.manager [-] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Took 1.40 seconds to deallocate network for instance. [ 1137.696665] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1137.696665] env[62814]: value = "task-4294541" [ 1137.696665] env[62814]: _type = "Task" [ 1137.696665] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.708867] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294541, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.838476] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d1e97a-391e-0842-ac9b-3d85bc9b1b7e, 'name': SearchDatastore_Task, 'duration_secs': 0.055682} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.841401] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1137.841637] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 33ab2d9e-e960-40ae-acf9-4fea0d9c830e/33ab2d9e-e960-40ae-acf9-4fea0d9c830e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1137.842121] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-baf97cc1-b8d3-4d25-b1d5-a02f7b86d158 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.851292] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1137.851292] env[62814]: value = "task-4294542" [ 1137.851292] env[62814]: _type = "Task" [ 1137.851292] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.862414] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294542, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.892488] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d25f4d-8c52-4269-bc3c-2d976a7d9ab8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.901078] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99309a5f-724f-49ef-986e-d9b11d230e51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.935344] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d8de2b6-e1a1-4777-ad50-8a7aed1fb239 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.944865] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0631b5-fc9c-4b6a-8ace-44bc32a167e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.951515] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1137.959153] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a28b15e8-4313-42be-a5ae-b4fb0c4299b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.961176] env[62814]: DEBUG nova.compute.provider_tree [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1137.969664] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1137.969664] env[62814]: value = "task-4294543" [ 1137.969664] env[62814]: _type = "Task" [ 1137.969664] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.980083] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294543, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.194285] env[62814]: DEBUG nova.compute.manager [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1138.194596] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1138.195655] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f412f6-0ed4-4029-9ace-6bf31f02234e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.200166] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1138.206950] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1138.208283] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-baa196ba-8b99-431f-a89d-8e8c7c0ef506 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.213905] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294541, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.215378] env[62814]: DEBUG oslo_vmware.api [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1138.215378] env[62814]: value = "task-4294544" [ 1138.215378] env[62814]: _type = "Task" [ 1138.215378] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.224354] env[62814]: DEBUG oslo_vmware.api [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294544, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.366014] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294542, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.466069] env[62814]: DEBUG nova.scheduler.client.report [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1138.482660] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1138.482927] env[62814]: DEBUG nova.compute.manager [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1138.483780] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bdd9c93-fecd-4b3f-8c92-7346a89599fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.491953] env[62814]: DEBUG oslo_concurrency.lockutils [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1138.491953] env[62814]: DEBUG oslo_concurrency.lockutils [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1138.491953] env[62814]: DEBUG nova.network.neutron [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1138.711705] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294541, 'name': ReconfigVM_Task, 'duration_secs': 0.799957} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.712016] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfigured VM instance instance-00000067 to attach disk [datastore2] b1d88997-e52f-41bd-b1b4-dd096d20d60a/b1d88997-e52f-41bd-b1b4-dd096d20d60a.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1138.712723] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f67c623c-fd35-43a7-bb88-1908625e5a92 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.720031] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1138.720031] env[62814]: value = "task-4294545" [ 1138.720031] env[62814]: _type = "Task" [ 1138.720031] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.725875] env[62814]: DEBUG oslo_vmware.api [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294544, 'name': PowerOffVM_Task, 'duration_secs': 0.315915} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.726456] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1138.726621] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1138.726854] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0a5a6a2c-48fe-4b23-9221-575e4bb85d32 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.731486] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294545, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.803187] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1138.803461] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1138.803644] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleting the datastore file [datastore2] eeb0255e-8c94-4564-af87-818cc8dee58e {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1138.803928] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-64fb7d9f-2385-4165-a57e-00155963cf2d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.812118] env[62814]: DEBUG oslo_vmware.api [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1138.812118] env[62814]: value = "task-4294547" [ 1138.812118] env[62814]: _type = "Task" [ 1138.812118] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.820810] env[62814]: DEBUG oslo_vmware.api [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294547, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.863073] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294542, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.820793} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.863361] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 33ab2d9e-e960-40ae-acf9-4fea0d9c830e/33ab2d9e-e960-40ae-acf9-4fea0d9c830e.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1138.863579] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1138.863865] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3433d585-954a-4640-ab0b-ec7c4731d487 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.873594] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1138.873594] env[62814]: value = "task-4294548" [ 1138.873594] env[62814]: _type = "Task" [ 1138.873594] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.882076] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294548, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.970535] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1138.971386] env[62814]: DEBUG nova.compute.manager [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1138.974796] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.699s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1138.975119] env[62814]: DEBUG nova.objects.instance [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lazy-loading 'resources' on Instance uuid d5df1d0e-55e6-452e-882a-053a83250b39 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1139.061140] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "53e13866-0174-4867-9fd9-2ea3f5ee399d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1139.062141] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "53e13866-0174-4867-9fd9-2ea3f5ee399d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1139.212279] env[62814]: DEBUG nova.network.neutron [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updating instance_info_cache with network_info: [{"id": "9950de47-d55a-480b-b88e-818e09b9bc80", "address": "fa:16:3e:0c:71:2f", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9950de47-d5", "ovs_interfaceid": "9950de47-d55a-480b-b88e-818e09b9bc80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.230722] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294545, 'name': Rename_Task, 'duration_secs': 0.200095} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.230996] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1139.231275] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51703c0e-770f-4dfb-9d1e-428337f0f04f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.239151] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1139.239151] env[62814]: value = "task-4294549" [ 1139.239151] env[62814]: _type = "Task" [ 1139.239151] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.249145] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294549, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.322037] env[62814]: DEBUG oslo_vmware.api [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294547, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.160557} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.322324] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1139.322547] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1139.322736] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1139.322990] env[62814]: INFO nova.compute.manager [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1139.323341] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1139.323469] env[62814]: DEBUG nova.compute.manager [-] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1139.323568] env[62814]: DEBUG nova.network.neutron [-] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1139.383162] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294548, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068597} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.383359] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1139.384201] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ff5a21-1620-49f7-ad6f-728194045cdb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.407723] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 33ab2d9e-e960-40ae-acf9-4fea0d9c830e/33ab2d9e-e960-40ae-acf9-4fea0d9c830e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1139.408074] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2bc1b8a5-0342-4185-910a-23aebdd826b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.427776] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1139.427776] env[62814]: value = "task-4294550" [ 1139.427776] env[62814]: _type = "Task" [ 1139.427776] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.436394] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294550, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.478240] env[62814]: DEBUG nova.compute.utils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1139.482960] env[62814]: DEBUG nova.compute.manager [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1139.483308] env[62814]: DEBUG nova.network.neutron [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1139.524074] env[62814]: DEBUG nova.policy [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51348be2493643fb9e5b44f27e5a7a94', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '513135674ee446d19fa8c667a47138a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1139.566153] env[62814]: DEBUG nova.compute.manager [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1139.716931] env[62814]: DEBUG oslo_concurrency.lockutils [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1139.751615] env[62814]: DEBUG oslo_vmware.api [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294549, 'name': PowerOnVM_Task, 'duration_secs': 0.496377} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.752120] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1139.752460] env[62814]: INFO nova.compute.manager [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Took 8.73 seconds to spawn the instance on the hypervisor. [ 1139.752687] env[62814]: DEBUG nova.compute.manager [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1139.753646] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27289e60-ecaf-47d3-93f9-a146837cf5b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.786859] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a51e2e5-59d3-42f9-a9f1-263a2402e85c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.794472] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6deb58-857e-48cf-9441-7d5b6ca0ba4f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.831161] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36df5f13-e580-4754-a8e1-919b33929a4e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.839575] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c6fe4e-c27e-4dac-b360-c950bb02cd7d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.858283] env[62814]: DEBUG nova.compute.provider_tree [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.882640] env[62814]: DEBUG nova.compute.manager [req-cbf563f0-0cc7-4bbf-9592-f63593962937 req-edeee43c-289e-4f61-a35a-9bd98df026c4 service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Received event network-vif-deleted-78214f32-dc80-4625-b6a8-a5f4d244cfb1 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1139.882969] env[62814]: INFO nova.compute.manager [req-cbf563f0-0cc7-4bbf-9592-f63593962937 req-edeee43c-289e-4f61-a35a-9bd98df026c4 service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Neutron deleted interface 78214f32-dc80-4625-b6a8-a5f4d244cfb1; detaching it from the instance and deleting it from the info cache [ 1139.883267] env[62814]: DEBUG nova.network.neutron [req-cbf563f0-0cc7-4bbf-9592-f63593962937 req-edeee43c-289e-4f61-a35a-9bd98df026c4 service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1139.939230] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294550, 'name': ReconfigVM_Task, 'duration_secs': 0.273159} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.939552] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 33ab2d9e-e960-40ae-acf9-4fea0d9c830e/33ab2d9e-e960-40ae-acf9-4fea0d9c830e.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1139.940227] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a55de7d4-00d3-461f-8fe9-001c9608fd84 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.946729] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1139.946729] env[62814]: value = "task-4294551" [ 1139.946729] env[62814]: _type = "Task" [ 1139.946729] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.955356] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294551, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.984043] env[62814]: DEBUG nova.compute.manager [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1140.088925] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.188288] env[62814]: DEBUG nova.network.neutron [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Successfully created port: 837d03b5-250f-44d2-bafb-31184267117d {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1140.278411] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1140.280842] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc72571-de4b-4850-96e1-124d3caeaaab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.288151] env[62814]: INFO nova.compute.manager [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Took 34.75 seconds to build instance. [ 1140.292378] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1140.292662] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64f72abd-4f48-4fba-b761-a18317e4c6dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.355464] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1140.355796] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1140.356033] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleting the datastore file [datastore2] edcdb1e5-ed9d-49e6-97e6-ea7629682547 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1140.356318] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc5d2e2b-621e-42b5-83f5-cc7525171b3e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.361472] env[62814]: DEBUG nova.scheduler.client.report [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1140.366740] env[62814]: DEBUG nova.network.neutron [-] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1140.369455] env[62814]: DEBUG nova.compute.manager [req-f2041392-e0c5-4154-9a4b-48784bc6764f req-28391696-e971-429a-802a-7d6c71e4f5c4 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Received event network-vif-unplugged-9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1140.370362] env[62814]: DEBUG oslo_concurrency.lockutils [req-f2041392-e0c5-4154-9a4b-48784bc6764f req-28391696-e971-429a-802a-7d6c71e4f5c4 service nova] Acquiring lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1140.370555] env[62814]: DEBUG oslo_concurrency.lockutils [req-f2041392-e0c5-4154-9a4b-48784bc6764f req-28391696-e971-429a-802a-7d6c71e4f5c4 service nova] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1140.370735] env[62814]: DEBUG oslo_concurrency.lockutils [req-f2041392-e0c5-4154-9a4b-48784bc6764f req-28391696-e971-429a-802a-7d6c71e4f5c4 service nova] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1140.370907] env[62814]: DEBUG nova.compute.manager [req-f2041392-e0c5-4154-9a4b-48784bc6764f req-28391696-e971-429a-802a-7d6c71e4f5c4 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] No waiting events found dispatching network-vif-unplugged-9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1140.371098] env[62814]: WARNING nova.compute.manager [req-f2041392-e0c5-4154-9a4b-48784bc6764f req-28391696-e971-429a-802a-7d6c71e4f5c4 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Received unexpected event network-vif-unplugged-9950de47-d55a-480b-b88e-818e09b9bc80 for instance with vm_state shelved and task_state shelving_offloading. [ 1140.371740] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1140.371740] env[62814]: value = "task-4294553" [ 1140.371740] env[62814]: _type = "Task" [ 1140.371740] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.382289] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294553, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.386383] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-181c31fc-7323-4318-a0a3-d59d03869ec7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.396073] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd8275b-5b5a-4038-ad60-4c00adf4b9af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.436744] env[62814]: DEBUG nova.compute.manager [req-cbf563f0-0cc7-4bbf-9592-f63593962937 req-edeee43c-289e-4f61-a35a-9bd98df026c4 service nova] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Detach interface failed, port_id=78214f32-dc80-4625-b6a8-a5f4d244cfb1, reason: Instance eeb0255e-8c94-4564-af87-818cc8dee58e could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1140.455768] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294551, 'name': Rename_Task, 'duration_secs': 0.143148} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.456343] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1140.456343] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25c61b5f-03f1-4099-8aa9-631988cd30f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.462760] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1140.462760] env[62814]: value = "task-4294554" [ 1140.462760] env[62814]: _type = "Task" [ 1140.462760] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.473267] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294554, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.789250] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e05995f6-f693-435f-ad2a-4b57211e1a1e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.261s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1140.875058] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.900s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1140.877508] env[62814]: INFO nova.compute.manager [-] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Took 1.55 seconds to deallocate network for instance. [ 1140.878023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.076s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1140.879094] env[62814]: DEBUG nova.objects.instance [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lazy-loading 'resources' on Instance uuid cfab341e-57a1-48b4-9b6b-ceecf28c223b {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1140.897463] env[62814]: DEBUG oslo_vmware.api [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294553, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127577} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.897778] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1140.897960] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1140.898150] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1140.940031] env[62814]: INFO nova.scheduler.client.report [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleted allocations for instance edcdb1e5-ed9d-49e6-97e6-ea7629682547 [ 1140.972732] env[62814]: DEBUG oslo_vmware.api [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294554, 'name': PowerOnVM_Task, 'duration_secs': 0.469209} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.973698] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1140.973904] env[62814]: INFO nova.compute.manager [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Took 7.57 seconds to spawn the instance on the hypervisor. [ 1140.974151] env[62814]: DEBUG nova.compute.manager [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1140.974876] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f8a881-a5da-448f-8394-2c7e629a41d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.993146] env[62814]: DEBUG nova.compute.manager [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1141.019578] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1141.019832] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1141.019990] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1141.020194] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1141.020348] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1141.020542] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1141.020808] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1141.020990] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1141.021186] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1141.021355] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1141.021530] env[62814]: DEBUG nova.virt.hardware [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1141.022669] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373e1c37-744b-4cbe-887e-5be6fbefefa0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.031801] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da26badb-4830-45b4-9ad2-2ae4b497db42 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.395295] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.401023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bb62754-ca46-4bdf-9d11-0fde30766723 tempest-ServerActionsV293TestJSON-372620725 tempest-ServerActionsV293TestJSON-372620725-project-member] Lock "d5df1d0e-55e6-452e-882a-053a83250b39" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.536s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1141.447194] env[62814]: DEBUG oslo_concurrency.lockutils [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.495894] env[62814]: INFO nova.compute.manager [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Took 31.71 seconds to build instance. [ 1141.629949] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e1d7fd-4ec1-4ace-a081-c0c537ce2d56 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.638171] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7bf89d9-17bb-4d9a-ad5f-730fda20592d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.672763] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9c1634-7adb-4cd0-ac30-345b1a2b653e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.681055] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6ff1742-ba52-431a-b645-5c70efde68fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.696443] env[62814]: DEBUG nova.compute.provider_tree [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.819458] env[62814]: DEBUG nova.compute.manager [req-da31eff7-bd4a-4328-b7ad-b0e4ad60d2ae req-ea13cc2f-7bab-4923-b467-8dfc3fdec139 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Received event network-vif-plugged-837d03b5-250f-44d2-bafb-31184267117d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1141.819923] env[62814]: DEBUG oslo_concurrency.lockutils [req-da31eff7-bd4a-4328-b7ad-b0e4ad60d2ae req-ea13cc2f-7bab-4923-b467-8dfc3fdec139 service nova] Acquiring lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1141.819923] env[62814]: DEBUG oslo_concurrency.lockutils [req-da31eff7-bd4a-4328-b7ad-b0e4ad60d2ae req-ea13cc2f-7bab-4923-b467-8dfc3fdec139 service nova] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1141.821079] env[62814]: DEBUG oslo_concurrency.lockutils [req-da31eff7-bd4a-4328-b7ad-b0e4ad60d2ae req-ea13cc2f-7bab-4923-b467-8dfc3fdec139 service nova] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1141.821079] env[62814]: DEBUG nova.compute.manager [req-da31eff7-bd4a-4328-b7ad-b0e4ad60d2ae req-ea13cc2f-7bab-4923-b467-8dfc3fdec139 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] No waiting events found dispatching network-vif-plugged-837d03b5-250f-44d2-bafb-31184267117d {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1141.821220] env[62814]: WARNING nova.compute.manager [req-da31eff7-bd4a-4328-b7ad-b0e4ad60d2ae req-ea13cc2f-7bab-4923-b467-8dfc3fdec139 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Received unexpected event network-vif-plugged-837d03b5-250f-44d2-bafb-31184267117d for instance with vm_state building and task_state spawning. [ 1141.908231] env[62814]: DEBUG nova.compute.manager [req-8430cd68-a0b0-43cb-9478-3c85274e348f req-943da967-6435-474c-8f9e-dba45a2a958e service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Received event network-changed-0235e89b-ec0d-4439-817c-dd76d6987b80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1141.908414] env[62814]: DEBUG nova.compute.manager [req-8430cd68-a0b0-43cb-9478-3c85274e348f req-943da967-6435-474c-8f9e-dba45a2a958e service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Refreshing instance network info cache due to event network-changed-0235e89b-ec0d-4439-817c-dd76d6987b80. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1141.908627] env[62814]: DEBUG oslo_concurrency.lockutils [req-8430cd68-a0b0-43cb-9478-3c85274e348f req-943da967-6435-474c-8f9e-dba45a2a958e service nova] Acquiring lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1141.908771] env[62814]: DEBUG oslo_concurrency.lockutils [req-8430cd68-a0b0-43cb-9478-3c85274e348f req-943da967-6435-474c-8f9e-dba45a2a958e service nova] Acquired lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1141.908931] env[62814]: DEBUG nova.network.neutron [req-8430cd68-a0b0-43cb-9478-3c85274e348f req-943da967-6435-474c-8f9e-dba45a2a958e service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Refreshing network info cache for port 0235e89b-ec0d-4439-817c-dd76d6987b80 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1141.997807] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f80b8c08-da55-4fb8-b525-f1f62735c015 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.223s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.200487] env[62814]: DEBUG nova.scheduler.client.report [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1142.400720] env[62814]: DEBUG nova.compute.manager [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Received event network-changed-9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1142.400996] env[62814]: DEBUG nova.compute.manager [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Refreshing instance network info cache due to event network-changed-9950de47-d55a-480b-b88e-818e09b9bc80. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1142.402247] env[62814]: DEBUG oslo_concurrency.lockutils [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] Acquiring lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.402422] env[62814]: DEBUG oslo_concurrency.lockutils [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] Acquired lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1142.402621] env[62814]: DEBUG nova.network.neutron [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Refreshing network info cache for port 9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1142.476291] env[62814]: DEBUG nova.network.neutron [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Successfully updated port: 837d03b5-250f-44d2-bafb-31184267117d {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1142.705314] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1142.707521] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.611s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1142.730474] env[62814]: INFO nova.scheduler.client.report [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Deleted allocations for instance cfab341e-57a1-48b4-9b6b-ceecf28c223b [ 1142.785542] env[62814]: DEBUG nova.network.neutron [req-8430cd68-a0b0-43cb-9478-3c85274e348f req-943da967-6435-474c-8f9e-dba45a2a958e service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updated VIF entry in instance network info cache for port 0235e89b-ec0d-4439-817c-dd76d6987b80. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1142.785935] env[62814]: DEBUG nova.network.neutron [req-8430cd68-a0b0-43cb-9478-3c85274e348f req-943da967-6435-474c-8f9e-dba45a2a958e service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance_info_cache with network_info: [{"id": "0235e89b-ec0d-4439-817c-dd76d6987b80", "address": "fa:16:3e:46:9d:6d", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0235e89b-ec", "ovs_interfaceid": "0235e89b-ec0d-4439-817c-dd76d6987b80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1142.979426] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.979702] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1142.979960] env[62814]: DEBUG nova.network.neutron [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1143.114802] env[62814]: DEBUG nova.network.neutron [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updated VIF entry in instance network info cache for port 9950de47-d55a-480b-b88e-818e09b9bc80. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1143.115283] env[62814]: DEBUG nova.network.neutron [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updating instance_info_cache with network_info: [{"id": "9950de47-d55a-480b-b88e-818e09b9bc80", "address": "fa:16:3e:0c:71:2f", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": null, "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9950de47-d5", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.241605] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5404b8d1-1acb-407b-89e5-0f16216312b5 tempest-ListServerFiltersTestJSON-1667271948 tempest-ListServerFiltersTestJSON-1667271948-project-member] Lock "cfab341e-57a1-48b4-9b6b-ceecf28c223b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.063s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1143.289049] env[62814]: DEBUG oslo_concurrency.lockutils [req-8430cd68-a0b0-43cb-9478-3c85274e348f req-943da967-6435-474c-8f9e-dba45a2a958e service nova] Releasing lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1143.515870] env[62814]: DEBUG nova.network.neutron [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1143.551399] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1143.618623] env[62814]: DEBUG oslo_concurrency.lockutils [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] Releasing lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1143.619066] env[62814]: DEBUG nova.compute.manager [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Received event network-changed-446b3cd1-e546-440b-bf99-c5bfb3dd4878 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1143.619342] env[62814]: DEBUG nova.compute.manager [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Refreshing instance network info cache due to event network-changed-446b3cd1-e546-440b-bf99-c5bfb3dd4878. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1143.619663] env[62814]: DEBUG oslo_concurrency.lockutils [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] Acquiring lock "refresh_cache-33ab2d9e-e960-40ae-acf9-4fea0d9c830e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.619894] env[62814]: DEBUG oslo_concurrency.lockutils [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] Acquired lock "refresh_cache-33ab2d9e-e960-40ae-acf9-4fea0d9c830e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1143.620154] env[62814]: DEBUG nova.network.neutron [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Refreshing network info cache for port 446b3cd1-e546-440b-bf99-c5bfb3dd4878 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1143.726483] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Applying migration context for instance 7e849b23-226d-4a1d-a85c-cf96964ce034 as it has an incoming, in-progress migration ef6d923a-0f7c-438d-bc3f-962f524c5f40. Migration status is confirming {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1143.726483] env[62814]: INFO nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating resource usage from migration ef6d923a-0f7c-438d-bc3f-962f524c5f40 [ 1143.731639] env[62814]: DEBUG nova.network.neutron [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance_info_cache with network_info: [{"id": "837d03b5-250f-44d2-bafb-31184267117d", "address": "fa:16:3e:24:ee:bf", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap837d03b5-25", "ovs_interfaceid": "837d03b5-250f-44d2-bafb-31184267117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.754489] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 7350d352-9336-40b8-81a6-0a4795d9f8dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1143.754678] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 0c6c598f-ef5a-4e91-b811-cd3d8f072647 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1143.754828] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 201cf7be-f81f-4eba-ae65-a370ad8e48c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1143.754965] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 1f13051e-e51b-4981-9445-d5420c3c1818 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1143.755113] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance f00d414e-1dfd-43cf-9245-7cbeea8850b3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1143.755725] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance e75c8039-9ff8-45f2-8770-894f2307cc14 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1143.755725] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Migration ef6d923a-0f7c-438d-bc3f-962f524c5f40 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1143.755725] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 7e849b23-226d-4a1d-a85c-cf96964ce034 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1143.755725] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 2d7a18f3-456b-470c-a759-632e5d79a8f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1143.755725] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance eeb0255e-8c94-4564-af87-818cc8dee58e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1143.755929] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance b1d88997-e52f-41bd-b1b4-dd096d20d60a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1143.755929] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 33ab2d9e-e960-40ae-acf9-4fea0d9c830e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1143.756037] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 606e31eb-2349-427f-9c9b-ed9dc5b385f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1144.237780] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1144.238191] env[62814]: DEBUG nova.compute.manager [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Instance network_info: |[{"id": "837d03b5-250f-44d2-bafb-31184267117d", "address": "fa:16:3e:24:ee:bf", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap837d03b5-25", "ovs_interfaceid": "837d03b5-250f-44d2-bafb-31184267117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1144.238619] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:ee:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '837d03b5-250f-44d2-bafb-31184267117d', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1144.249968] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1144.250970] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1144.250970] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c18c75d-b007-4a10-85c3-26935805dc4c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.269532] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d65ccde2-b31a-4032-8795-1d609fdfcc73 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1144.277799] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1144.277799] env[62814]: value = "task-4294555" [ 1144.277799] env[62814]: _type = "Task" [ 1144.277799] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.286851] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294555, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.379157] env[62814]: DEBUG nova.network.neutron [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Updated VIF entry in instance network info cache for port 446b3cd1-e546-440b-bf99-c5bfb3dd4878. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1144.379801] env[62814]: DEBUG nova.network.neutron [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Updating instance_info_cache with network_info: [{"id": "446b3cd1-e546-440b-bf99-c5bfb3dd4878", "address": "fa:16:3e:d1:71:e0", "network": {"id": "5363651c-6ade-4a4c-b762-9fc7cbbcf5ea", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1833251571-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.176", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2b3ba1c9c7624abfb3e0bd880dc40166", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "50cf0a70-948d-4611-af05-94c1483064ed", "external-id": "nsx-vlan-transportzone-536", "segmentation_id": 536, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap446b3cd1-e5", "ovs_interfaceid": "446b3cd1-e546-440b-bf99-c5bfb3dd4878", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.439988] env[62814]: DEBUG nova.compute.manager [req-c1f214e3-5d9f-40cc-a9c1-2421ab0c4686 req-7d993b3a-0751-4f24-a91f-d81da4688248 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Received event network-changed-837d03b5-250f-44d2-bafb-31184267117d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1144.440277] env[62814]: DEBUG nova.compute.manager [req-c1f214e3-5d9f-40cc-a9c1-2421ab0c4686 req-7d993b3a-0751-4f24-a91f-d81da4688248 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Refreshing instance network info cache due to event network-changed-837d03b5-250f-44d2-bafb-31184267117d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1144.440426] env[62814]: DEBUG oslo_concurrency.lockutils [req-c1f214e3-5d9f-40cc-a9c1-2421ab0c4686 req-7d993b3a-0751-4f24-a91f-d81da4688248 service nova] Acquiring lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.440572] env[62814]: DEBUG oslo_concurrency.lockutils [req-c1f214e3-5d9f-40cc-a9c1-2421ab0c4686 req-7d993b3a-0751-4f24-a91f-d81da4688248 service nova] Acquired lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1144.440742] env[62814]: DEBUG nova.network.neutron [req-c1f214e3-5d9f-40cc-a9c1-2421ab0c4686 req-7d993b3a-0751-4f24-a91f-d81da4688248 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Refreshing network info cache for port 837d03b5-250f-44d2-bafb-31184267117d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1144.773108] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 29468dc9-4254-4ac6-989c-c7fa0c4cc682 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1144.790968] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294555, 'name': CreateVM_Task, 'duration_secs': 0.508925} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.791208] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1144.791900] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.792082] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1144.792402] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1144.792897] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb4ef968-72e0-4860-ad36-6ddde79a8592 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.798152] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1144.798152] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dac4e0-6a74-160c-139e-e728071288fd" [ 1144.798152] env[62814]: _type = "Task" [ 1144.798152] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.807480] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dac4e0-6a74-160c-139e-e728071288fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.883214] env[62814]: DEBUG oslo_concurrency.lockutils [req-9d342986-1281-44e8-8f5d-dbee75860832 req-78d6557d-40d6-4088-b938-3c7129d9ca86 service nova] Releasing lock "refresh_cache-33ab2d9e-e960-40ae-acf9-4fea0d9c830e" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.182639] env[62814]: DEBUG nova.network.neutron [req-c1f214e3-5d9f-40cc-a9c1-2421ab0c4686 req-7d993b3a-0751-4f24-a91f-d81da4688248 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updated VIF entry in instance network info cache for port 837d03b5-250f-44d2-bafb-31184267117d. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1145.183103] env[62814]: DEBUG nova.network.neutron [req-c1f214e3-5d9f-40cc-a9c1-2421ab0c4686 req-7d993b3a-0751-4f24-a91f-d81da4688248 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance_info_cache with network_info: [{"id": "837d03b5-250f-44d2-bafb-31184267117d", "address": "fa:16:3e:24:ee:bf", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap837d03b5-25", "ovs_interfaceid": "837d03b5-250f-44d2-bafb-31184267117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.275681] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 53e13866-0174-4867-9fd9-2ea3f5ee399d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1145.275861] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (7136a6f7-3927-4aa1-a4e9-7fcbd9976745): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 1145.276103] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 10 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1145.276252] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2496MB phys_disk=149GB used_disk=10GB total_vcpus=48 used_vcpus=10 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1145.308386] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dac4e0-6a74-160c-139e-e728071288fd, 'name': SearchDatastore_Task, 'duration_secs': 0.010715} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.308689] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.308935] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1145.309157] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.309308] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1145.309488] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1145.309740] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af22ec7f-29c9-4e59-a4e2-8ec60fa4dc7e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.320233] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1145.320405] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1145.321111] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edcee5a1-9164-4462-a0ff-da75e88b4aad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.328801] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1145.328801] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5271eca5-02c7-0f79-3085-9828eb58e4ab" [ 1145.328801] env[62814]: _type = "Task" [ 1145.328801] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.337992] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5271eca5-02c7-0f79-3085-9828eb58e4ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.469666] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31df86b-98c7-4c95-9dca-920725905216 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.477561] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc5f5978-8654-4aa2-b762-1c4355c7a6e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.506911] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4202d911-c040-438c-81e5-23f9b9dc3cd3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.514233] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee975475-dfc4-434e-9fa5-e9a213bc8513 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.526905] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.686585] env[62814]: DEBUG oslo_concurrency.lockutils [req-c1f214e3-5d9f-40cc-a9c1-2421ab0c4686 req-7d993b3a-0751-4f24-a91f-d81da4688248 service nova] Releasing lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1145.841016] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5271eca5-02c7-0f79-3085-9828eb58e4ab, 'name': SearchDatastore_Task, 'duration_secs': 0.008657} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.841961] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f88f0bb-22fa-4773-8655-9d920508ced0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.846980] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1145.846980] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5231bf2b-d406-e1a0-01f6-9c5b9092cff9" [ 1145.846980] env[62814]: _type = "Task" [ 1145.846980] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.854638] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5231bf2b-d406-e1a0-01f6-9c5b9092cff9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.030356] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1146.357903] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5231bf2b-d406-e1a0-01f6-9c5b9092cff9, 'name': SearchDatastore_Task, 'duration_secs': 0.009668} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.358187] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1146.358452] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 606e31eb-2349-427f-9c9b-ed9dc5b385f5/606e31eb-2349-427f-9c9b-ed9dc5b385f5.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1146.358696] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fd15aa26-2a06-43f2-b161-e8808dd49a65 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.365472] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1146.365472] env[62814]: value = "task-4294556" [ 1146.365472] env[62814]: _type = "Task" [ 1146.365472] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.372271] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294556, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.541974] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1146.542246] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.835s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1146.542526] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.980s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1146.544227] env[62814]: INFO nova.compute.claims [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1146.877445] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294556, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.377048] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294556, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.649097} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.377404] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 606e31eb-2349-427f-9c9b-ed9dc5b385f5/606e31eb-2349-427f-9c9b-ed9dc5b385f5.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1147.377629] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1147.377885] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da0b476a-7f0a-409a-b445-b784d342eb04 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.385654] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1147.385654] env[62814]: value = "task-4294557" [ 1147.385654] env[62814]: _type = "Task" [ 1147.385654] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.395619] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294557, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.732370] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52162940-acdf-4b24-b11e-47fe2e9189f3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.740279] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e48aa8-60ce-4820-848b-efdc4596bffa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.771762] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983649a2-e108-40fc-adaa-c382efac6e6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.779296] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af201ecb-eb86-4637-9013-d9496e9351b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.792690] env[62814]: DEBUG nova.compute.provider_tree [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1147.894860] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294557, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075393} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.895225] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1147.895945] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be84dd30-0a4b-4acc-97f1-79b8bbbe43fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.917429] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 606e31eb-2349-427f-9c9b-ed9dc5b385f5/606e31eb-2349-427f-9c9b-ed9dc5b385f5.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1147.918128] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23a8e5e6-dfe4-486b-96b4-12f477f394cf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.937906] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1147.937906] env[62814]: value = "task-4294558" [ 1147.937906] env[62814]: _type = "Task" [ 1147.937906] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.947272] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294558, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.295952] env[62814]: DEBUG nova.scheduler.client.report [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1148.448719] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294558, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.800700] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1148.801218] env[62814]: DEBUG nova.compute.manager [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1148.804117] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.028s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1148.805555] env[62814]: INFO nova.compute.claims [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1148.948437] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294558, 'name': ReconfigVM_Task, 'duration_secs': 0.760046} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.948437] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 606e31eb-2349-427f-9c9b-ed9dc5b385f5/606e31eb-2349-427f-9c9b-ed9dc5b385f5.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1148.949503] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cac81f91-e769-4be7-898f-e18f0ab43254 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.955931] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1148.955931] env[62814]: value = "task-4294559" [ 1148.955931] env[62814]: _type = "Task" [ 1148.955931] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.964034] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294559, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.310649] env[62814]: DEBUG nova.compute.utils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1149.315063] env[62814]: DEBUG nova.compute.manager [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Not allocating networking since 'none' was specified. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 1149.466891] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294559, 'name': Rename_Task, 'duration_secs': 0.1453} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.467896] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1149.467896] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-adeb6c84-4a21-4352-8267-a56bee411829 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.474116] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1149.474116] env[62814]: value = "task-4294560" [ 1149.474116] env[62814]: _type = "Task" [ 1149.474116] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.485709] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294560, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.815093] env[62814]: DEBUG nova.compute.manager [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1149.983998] env[62814]: DEBUG oslo_vmware.api [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294560, 'name': PowerOnVM_Task, 'duration_secs': 0.4801} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.986496] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1149.986701] env[62814]: INFO nova.compute.manager [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Took 8.99 seconds to spawn the instance on the hypervisor. [ 1149.986881] env[62814]: DEBUG nova.compute.manager [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1149.988103] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acdb802-b469-4fe7-8d99-e31125f9a30a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.060127] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c64e3a-9c86-4ed4-a3ff-e003fc2d9356 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.067966] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd74a2be-cb45-4355-a9e6-7bb53d99129b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.097901] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56080602-28ac-4d2a-9a8f-9ba92e353f52 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.105974] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a64c8b42-b89b-413b-97d7-93c69326650a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.119965] env[62814]: DEBUG nova.compute.provider_tree [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1150.507977] env[62814]: INFO nova.compute.manager [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Took 27.36 seconds to build instance. [ 1150.623393] env[62814]: DEBUG nova.scheduler.client.report [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1150.827483] env[62814]: DEBUG nova.compute.manager [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1150.858360] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1150.858498] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1150.858653] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1150.858832] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1150.858978] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1150.859166] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1150.859382] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1150.859559] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1150.859726] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1150.859886] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1150.860071] env[62814]: DEBUG nova.virt.hardware [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1150.860998] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720df90a-8b3c-43c7-8662-934e27442b0d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.869300] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9baa6a-cc8b-4724-81bd-509e02731890 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.883261] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1150.888961] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Creating folder: Project (ea13ca78a8d8451fa9067159fd81f250). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1150.889245] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7722ed1c-33db-49ec-8ea2-3de161d8c0b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.898765] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Created folder: Project (ea13ca78a8d8451fa9067159fd81f250) in parent group-v845547. [ 1150.898952] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Creating folder: Instances. Parent ref: group-v845837. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1150.899197] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f878215d-1108-4224-8734-a08392773d3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.908215] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Created folder: Instances in parent group-v845837. [ 1150.908440] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1150.908625] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1150.908818] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ace325c4-9e65-4a54-a5d5-ad55091d5b2f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.925757] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1150.925757] env[62814]: value = "task-4294563" [ 1150.925757] env[62814]: _type = "Task" [ 1150.925757] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.932854] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294563, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.010682] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6c7c56b7-b708-4495-b4e5-26441b420bee tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.878s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.129436] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.130122] env[62814]: DEBUG nova.compute.manager [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1151.133321] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.321s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.133861] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.135971] env[62814]: DEBUG oslo_concurrency.lockutils [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.484s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.136324] env[62814]: DEBUG oslo_concurrency.lockutils [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.143477] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 16.149s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.188457] env[62814]: INFO nova.scheduler.client.report [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Deleted allocations for instance e75c8039-9ff8-45f2-8770-894f2307cc14 [ 1151.282512] env[62814]: DEBUG nova.compute.manager [req-59d2f6ce-7c68-4949-9fcc-bf05c243cb1a req-7941e5c1-931f-4769-9a92-90d5afba679b service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Received event network-changed-837d03b5-250f-44d2-bafb-31184267117d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1151.282758] env[62814]: DEBUG nova.compute.manager [req-59d2f6ce-7c68-4949-9fcc-bf05c243cb1a req-7941e5c1-931f-4769-9a92-90d5afba679b service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Refreshing instance network info cache due to event network-changed-837d03b5-250f-44d2-bafb-31184267117d. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1151.283024] env[62814]: DEBUG oslo_concurrency.lockutils [req-59d2f6ce-7c68-4949-9fcc-bf05c243cb1a req-7941e5c1-931f-4769-9a92-90d5afba679b service nova] Acquiring lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.283228] env[62814]: DEBUG oslo_concurrency.lockutils [req-59d2f6ce-7c68-4949-9fcc-bf05c243cb1a req-7941e5c1-931f-4769-9a92-90d5afba679b service nova] Acquired lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.283441] env[62814]: DEBUG nova.network.neutron [req-59d2f6ce-7c68-4949-9fcc-bf05c243cb1a req-7941e5c1-931f-4769-9a92-90d5afba679b service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Refreshing network info cache for port 837d03b5-250f-44d2-bafb-31184267117d {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1151.435907] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294563, 'name': CreateVM_Task, 'duration_secs': 0.26082} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.436084] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1151.436517] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.436682] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.437028] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1151.437296] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55e69a93-7cd6-42f7-a944-5561dad560e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.441927] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1151.441927] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525116cf-ce1c-291b-8b67-5fcdcd4d9a62" [ 1151.441927] env[62814]: _type = "Task" [ 1151.441927] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.450143] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525116cf-ce1c-291b-8b67-5fcdcd4d9a62, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.642730] env[62814]: DEBUG nova.compute.utils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1151.646533] env[62814]: DEBUG nova.compute.manager [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Not allocating networking since 'none' was specified. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 1151.651691] env[62814]: DEBUG oslo_concurrency.lockutils [None req-32e49114-ac22-4e14-9f91-dea0878fea0b tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 38.877s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.653272] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 18.113s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.653564] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "8af98433-18eb-4a12-b18d-39b0a58821c0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1151.654035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1151.654279] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.656105] env[62814]: INFO nova.compute.manager [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Terminating instance [ 1151.698344] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b2b1ca51-02d6-401f-ab8e-784e0d3f1234 tempest-ServersNegativeTestMultiTenantJSON-1857537205 tempest-ServersNegativeTestMultiTenantJSON-1857537205-project-member] Lock "e75c8039-9ff8-45f2-8770-894f2307cc14" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.470s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1151.899564] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb59b56e-83b2-4489-a8d5-7272475d5860 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.907740] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89494b10-491b-4aef-923f-d9863a1a9977 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.940848] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99476806-8566-4f17-95d2-c52592aa8b2f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.954117] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290fcc73-bce4-4e97-b392-2f217735ae9c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.957850] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525116cf-ce1c-291b-8b67-5fcdcd4d9a62, 'name': SearchDatastore_Task, 'duration_secs': 0.03059} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.958158] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1151.958390] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1151.958618] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.958762] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1151.958940] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1151.959576] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d83f56f2-cbf9-4141-826f-bda3693f67b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.968633] env[62814]: DEBUG nova.compute.provider_tree [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1151.978510] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1151.978694] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1151.979435] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef677e92-a292-43c3-a421-dcbe7b4ba01f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.984433] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1151.984433] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5263a751-b31c-4e19-11c6-e3b6123d0af7" [ 1151.984433] env[62814]: _type = "Task" [ 1151.984433] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.992125] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5263a751-b31c-4e19-11c6-e3b6123d0af7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.059578] env[62814]: DEBUG nova.network.neutron [req-59d2f6ce-7c68-4949-9fcc-bf05c243cb1a req-7941e5c1-931f-4769-9a92-90d5afba679b service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updated VIF entry in instance network info cache for port 837d03b5-250f-44d2-bafb-31184267117d. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1152.059958] env[62814]: DEBUG nova.network.neutron [req-59d2f6ce-7c68-4949-9fcc-bf05c243cb1a req-7941e5c1-931f-4769-9a92-90d5afba679b service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance_info_cache with network_info: [{"id": "837d03b5-250f-44d2-bafb-31184267117d", "address": "fa:16:3e:24:ee:bf", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap837d03b5-25", "ovs_interfaceid": "837d03b5-250f-44d2-bafb-31184267117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.147232] env[62814]: DEBUG nova.compute.manager [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1152.160723] env[62814]: DEBUG nova.compute.manager [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1152.161572] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1152.161572] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-53ccb4a0-1cf2-4237-a800-d2c292876b61 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.170940] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee59a81-5472-469c-955d-57ca06e3075f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.204937] env[62814]: WARNING nova.virt.vmwareapi.vmops [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8af98433-18eb-4a12-b18d-39b0a58821c0 could not be found. [ 1152.205178] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1152.205367] env[62814]: INFO nova.compute.manager [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 1152.205617] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1152.205903] env[62814]: DEBUG nova.compute.manager [-] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1152.205996] env[62814]: DEBUG nova.network.neutron [-] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1152.474538] env[62814]: DEBUG nova.scheduler.client.report [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1152.496885] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5263a751-b31c-4e19-11c6-e3b6123d0af7, 'name': SearchDatastore_Task, 'duration_secs': 0.037905} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.497859] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a666a42-6e0d-4b6b-890c-d6dad3a485f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.503263] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1152.503263] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52df1492-4810-7eec-1979-94c4830ccb58" [ 1152.503263] env[62814]: _type = "Task" [ 1152.503263] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.512182] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52df1492-4810-7eec-1979-94c4830ccb58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.563428] env[62814]: DEBUG oslo_concurrency.lockutils [req-59d2f6ce-7c68-4949-9fcc-bf05c243cb1a req-7941e5c1-931f-4769-9a92-90d5afba679b service nova] Releasing lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.014526] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52df1492-4810-7eec-1979-94c4830ccb58, 'name': SearchDatastore_Task, 'duration_secs': 0.010757} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.015678] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1153.015678] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d65ccde2-b31a-4032-8795-1d609fdfcc73/d65ccde2-b31a-4032-8795-1d609fdfcc73.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1153.015678] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-537940f5-0973-4895-9f96-04672ec4e891 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.022512] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1153.022512] env[62814]: value = "task-4294564" [ 1153.022512] env[62814]: _type = "Task" [ 1153.022512] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.029906] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294564, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.101488] env[62814]: DEBUG oslo_concurrency.lockutils [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "1f13051e-e51b-4981-9445-d5420c3c1818" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1153.101756] env[62814]: DEBUG oslo_concurrency.lockutils [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.101938] env[62814]: INFO nova.compute.manager [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Shelving [ 1153.160482] env[62814]: DEBUG nova.compute.manager [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1153.164720] env[62814]: DEBUG nova.network.neutron [-] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.191334] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1153.191587] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1153.191746] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1153.191930] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1153.192099] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1153.192257] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1153.192478] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1153.192638] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1153.192805] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1153.196024] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1153.196024] env[62814]: DEBUG nova.virt.hardware [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1153.196024] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff88c67d-1973-40d7-b35e-6f6e6d4fe783 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.202807] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6adf66-477f-407f-97b6-f6ce63b505d1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.219052] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1153.225545] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1153.225904] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1153.226146] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5eeb50c7-9c5f-45fa-8e00-6533249c9248 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.246911] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1153.246911] env[62814]: value = "task-4294565" [ 1153.246911] env[62814]: _type = "Task" [ 1153.246911] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.256772] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294565, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.485087] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.346s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.488508] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.288s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.492029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1153.492029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.402s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1153.492618] env[62814]: INFO nova.compute.claims [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1153.523969] env[62814]: INFO nova.scheduler.client.report [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Deleted allocations for instance 0c6c598f-ef5a-4e91-b811-cd3d8f072647 [ 1153.544525] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294564, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518424} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.544931] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d65ccde2-b31a-4032-8795-1d609fdfcc73/d65ccde2-b31a-4032-8795-1d609fdfcc73.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1153.545608] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1153.545993] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e187f317-27e8-4c28-9529-0949c615b238 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.554295] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1153.554295] env[62814]: value = "task-4294566" [ 1153.554295] env[62814]: _type = "Task" [ 1153.554295] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.563871] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294566, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.667041] env[62814]: INFO nova.compute.manager [-] [instance: 8af98433-18eb-4a12-b18d-39b0a58821c0] Took 1.46 seconds to deallocate network for instance. [ 1153.757757] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294565, 'name': CreateVM_Task, 'duration_secs': 0.468453} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.757935] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1153.758388] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1153.759029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1153.759029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1153.759173] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f5b1da5-0d81-4ef4-8ac8-980129bbb53e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.763840] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1153.763840] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520db53c-d6c4-4146-a992-6d4282444adf" [ 1153.763840] env[62814]: _type = "Task" [ 1153.763840] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.771624] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520db53c-d6c4-4146-a992-6d4282444adf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.040972] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf024d93-58b6-4dcf-9e7a-b964238106aa tempest-FloatingIPsAssociationTestJSON-1426094244 tempest-FloatingIPsAssociationTestJSON-1426094244-project-member] Lock "0c6c598f-ef5a-4e91-b811-cd3d8f072647" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.410s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.065961] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294566, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066068} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.066396] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1154.068201] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d41803-e7a2-40e1-b1f2-c73d92354aa5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.073454] env[62814]: INFO nova.scheduler.client.report [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted allocation for migration ef6d923a-0f7c-438d-bc3f-962f524c5f40 [ 1154.096345] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] d65ccde2-b31a-4032-8795-1d609fdfcc73/d65ccde2-b31a-4032-8795-1d609fdfcc73.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1154.096978] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0703bf45-8c30-4283-b2fd-4a6c50a429a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.120251] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1154.120251] env[62814]: value = "task-4294567" [ 1154.120251] env[62814]: _type = "Task" [ 1154.120251] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.125662] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1154.125791] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25538c9e-70a3-4ab5-a99f-26806284bf84 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.132596] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294567, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.133942] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1154.133942] env[62814]: value = "task-4294568" [ 1154.133942] env[62814]: _type = "Task" [ 1154.133942] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.142458] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294568, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.278984] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520db53c-d6c4-4146-a992-6d4282444adf, 'name': SearchDatastore_Task, 'duration_secs': 0.009285} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.278984] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1154.278984] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1154.278984] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1154.278984] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1154.279417] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1154.279417] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5ce4bec6-aa01-44ae-97a8-f785f20f9512 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.289941] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1154.289941] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1154.291038] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e92a465c-24dc-4684-bab1-060ac6c8b85d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.297409] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1154.297409] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52edbd4c-d3fd-0c80-d2f1-3150a6ae9da5" [ 1154.297409] env[62814]: _type = "Task" [ 1154.297409] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.308570] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52edbd4c-d3fd-0c80-d2f1-3150a6ae9da5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.580715] env[62814]: DEBUG oslo_concurrency.lockutils [None req-79f7731f-81fe-4cc5-8b4a-5866761a1020 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 23.148s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.647803] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294567, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.656757] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294568, 'name': PowerOffVM_Task, 'duration_secs': 0.219077} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.658183] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1154.658515] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6864219b-9731-400d-a820-ec2fb02f5a3d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.683059] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05695811-0e45-45d8-9287-449f82d47b1a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.703029] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0f005dee-5001-4da9-a0d0-8ab91cd0f324 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "8af98433-18eb-4a12-b18d-39b0a58821c0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.050s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1154.806469] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50df758b-6abb-49e4-86dc-005f45da68fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.816654] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40811878-ed91-43c5-b601-ca93188a1cef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.820208] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52edbd4c-d3fd-0c80-d2f1-3150a6ae9da5, 'name': SearchDatastore_Task, 'duration_secs': 0.016699} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.821383] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a14bcd46-3b15-4fac-b7e0-f3d770e4ad79 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.849427] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf9703b-45aa-45c3-8b4b-2d9a9b6d0294 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.854752] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1154.854752] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5242b219-2a68-8524-f3e3-1c2869f9529c" [ 1154.854752] env[62814]: _type = "Task" [ 1154.854752] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.862309] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa701a7e-da6b-4432-85f0-613bf448154c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.870921] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5242b219-2a68-8524-f3e3-1c2869f9529c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.881418] env[62814]: DEBUG nova.compute.provider_tree [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1155.136155] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294567, 'name': ReconfigVM_Task, 'duration_secs': 0.729058} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.136155] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Reconfigured VM instance instance-0000006a to attach disk [datastore2] d65ccde2-b31a-4032-8795-1d609fdfcc73/d65ccde2-b31a-4032-8795-1d609fdfcc73.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1155.136155] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc9a76f0-a7aa-43b2-9717-7034d89d7022 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.143302] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1155.143302] env[62814]: value = "task-4294569" [ 1155.143302] env[62814]: _type = "Task" [ 1155.143302] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.151165] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294569, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.197195] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1155.197195] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-e86180de-db16-4686-b936-065fa8da97d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.202751] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1155.202751] env[62814]: value = "task-4294570" [ 1155.202751] env[62814]: _type = "Task" [ 1155.202751] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.216033] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294570, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.367578] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5242b219-2a68-8524-f3e3-1c2869f9529c, 'name': SearchDatastore_Task, 'duration_secs': 0.013074} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.368031] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1155.368470] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 29468dc9-4254-4ac6-989c-c7fa0c4cc682/29468dc9-4254-4ac6-989c-c7fa0c4cc682.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1155.371023] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8217499-c531-4466-91ae-f88063435cf9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.377904] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1155.377904] env[62814]: value = "task-4294571" [ 1155.377904] env[62814]: _type = "Task" [ 1155.377904] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.388294] env[62814]: DEBUG nova.scheduler.client.report [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1155.393296] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294571, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.652624] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294569, 'name': Rename_Task, 'duration_secs': 0.193717} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.653303] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1155.653961] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e39a9a1d-d84e-4efa-bc16-51c696aeb2c2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.662620] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1155.662620] env[62814]: value = "task-4294572" [ 1155.662620] env[62814]: _type = "Task" [ 1155.662620] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.676292] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294572, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.721752] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294570, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.895574] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1155.896145] env[62814]: DEBUG nova.compute.manager [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1155.898985] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294571, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.899555] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.504s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1155.899747] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1155.902328] env[62814]: DEBUG oslo_concurrency.lockutils [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.458s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1155.902574] env[62814]: DEBUG oslo_concurrency.lockutils [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1155.933677] env[62814]: INFO nova.scheduler.client.report [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted allocations for instance eeb0255e-8c94-4564-af87-818cc8dee58e [ 1156.174935] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294572, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.219716] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294570, 'name': CreateSnapshot_Task, 'duration_secs': 0.92289} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.220148] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1156.221140] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbac8500-4480-4754-b80d-cc78d64ec165 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.347606] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "881e9481-a5a9-489b-8abd-8efdaa239a11" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.347786] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "881e9481-a5a9-489b-8abd-8efdaa239a11" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.391033] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294571, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575393} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.391399] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 29468dc9-4254-4ac6-989c-c7fa0c4cc682/29468dc9-4254-4ac6-989c-c7fa0c4cc682.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1156.393020] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1156.393020] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-17049dfd-cec2-4b8e-af3d-8354cd7c83dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.399355] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1156.399355] env[62814]: value = "task-4294573" [ 1156.399355] env[62814]: _type = "Task" [ 1156.399355] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.400909] env[62814]: DEBUG nova.compute.utils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1156.407947] env[62814]: DEBUG nova.compute.manager [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Not allocating networking since 'none' was specified. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1984}} [ 1156.418988] env[62814]: DEBUG oslo_concurrency.lockutils [None req-42fafd94-9da0-4dae-965a-2bb168d4e2a6 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 38.418s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.420589] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294573, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.420877] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 12.870s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.421180] env[62814]: INFO nova.compute.manager [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Unshelving [ 1156.446775] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3d5c974e-b794-410e-8024-4d4ea8b47340 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "eeb0255e-8c94-4564-af87-818cc8dee58e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.760s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.675601] env[62814]: DEBUG oslo_vmware.api [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294572, 'name': PowerOnVM_Task, 'duration_secs': 0.758402} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.675886] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1156.676157] env[62814]: INFO nova.compute.manager [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Took 5.85 seconds to spawn the instance on the hypervisor. [ 1156.676441] env[62814]: DEBUG nova.compute.manager [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1156.677288] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994c406a-99d6-4fbe-943a-b5c0824807d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.743669] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1156.748725] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-82b4863c-a36e-486e-a14d-b6bc012e4982 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.755803] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1156.755803] env[62814]: value = "task-4294574" [ 1156.755803] env[62814]: _type = "Task" [ 1156.755803] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1156.767272] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294574, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.818609] env[62814]: DEBUG oslo_concurrency.lockutils [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.818888] env[62814]: DEBUG oslo_concurrency.lockutils [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.819212] env[62814]: DEBUG oslo_concurrency.lockutils [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.819463] env[62814]: DEBUG oslo_concurrency.lockutils [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.819695] env[62814]: DEBUG oslo_concurrency.lockutils [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.821922] env[62814]: INFO nova.compute.manager [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Terminating instance [ 1156.851362] env[62814]: DEBUG nova.compute.manager [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1156.909049] env[62814]: DEBUG nova.compute.manager [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1156.919643] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294573, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.144396} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.920807] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1156.921711] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25e97d7-4377-4b50-a717-cdfed1a333bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.933331] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "7e849b23-226d-4a1d-a85c-cf96964ce034" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.934151] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.934151] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "7e849b23-226d-4a1d-a85c-cf96964ce034-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1156.934151] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1156.934322] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1156.942192] env[62814]: INFO nova.compute.manager [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Terminating instance [ 1156.968753] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 29468dc9-4254-4ac6-989c-c7fa0c4cc682/29468dc9-4254-4ac6-989c-c7fa0c4cc682.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1156.968753] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61476f7a-500d-47bc-8517-01adc1605711 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.995611] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1156.995611] env[62814]: value = "task-4294575" [ 1156.995611] env[62814]: _type = "Task" [ 1156.995611] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.008103] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294575, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.200639] env[62814]: INFO nova.compute.manager [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Took 27.67 seconds to build instance. [ 1157.268206] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294574, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.325724] env[62814]: DEBUG nova.compute.manager [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1157.325985] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1157.326943] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae20a14-c10f-41c8-a360-c9a44345de7e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.335416] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1157.335678] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54e621a0-5324-4778-82a0-fbd32085a3b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.342297] env[62814]: DEBUG oslo_vmware.api [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1157.342297] env[62814]: value = "task-4294576" [ 1157.342297] env[62814]: _type = "Task" [ 1157.342297] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.352951] env[62814]: DEBUG oslo_vmware.api [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294576, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.375709] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.376066] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.378071] env[62814]: INFO nova.compute.claims [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1157.458455] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.489031] env[62814]: DEBUG nova.compute.manager [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1157.489277] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1157.490467] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4816347e-8381-448b-af69-bb7665f866b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.502687] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1157.503442] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46d3681f-3453-4575-943d-6407ea0d1893 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.509472] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294575, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.513164] env[62814]: DEBUG oslo_vmware.api [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1157.513164] env[62814]: value = "task-4294577" [ 1157.513164] env[62814]: _type = "Task" [ 1157.513164] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.521919] env[62814]: DEBUG oslo_vmware.api [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294577, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.546633] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "40e3c00b-2129-476e-bc67-a0ef13781990" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1157.546825] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "40e3c00b-2129-476e-bc67-a0ef13781990" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1157.702644] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4797db88-9ebd-48b6-b2a9-13c915038494 tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "d65ccde2-b31a-4032-8795-1d609fdfcc73" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.184s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1157.767089] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294574, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.852186] env[62814]: DEBUG oslo_vmware.api [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294576, 'name': PowerOffVM_Task, 'duration_secs': 0.251212} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1157.852481] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1157.852736] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1157.853046] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-070f36e3-259a-4423-b13c-bfa3044e8774 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.912312] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1157.912559] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1157.912759] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleting the datastore file [datastore2] 201cf7be-f81f-4eba-ae65-a370ad8e48c3 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1157.913036] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-039b0579-341b-42cd-9ae6-b7a3ab81a7d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.920267] env[62814]: DEBUG oslo_vmware.api [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1157.920267] env[62814]: value = "task-4294579" [ 1157.920267] env[62814]: _type = "Task" [ 1157.920267] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.930399] env[62814]: DEBUG nova.compute.manager [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1157.931642] env[62814]: DEBUG oslo_vmware.api [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294579, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1157.956555] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1157.956744] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1157.956922] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1157.957165] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1157.957328] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1157.957477] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1157.957706] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1157.957881] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1157.958081] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1157.958272] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1157.958452] env[62814]: DEBUG nova.virt.hardware [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1157.959675] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2caa42f9-eeef-4a7c-ba1c-6aa90dfe5215 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.967999] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2fab908-45d6-4393-8ee3-b31cda2a87f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.985914] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1157.992017] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Creating folder: Project (7da7ad544b704286a8138bfe82586c6b). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1157.992017] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f453b332-2a4d-47b1-8d57-ea75ffa33a45 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.005288] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294575, 'name': ReconfigVM_Task, 'duration_secs': 0.589017} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.006635] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 29468dc9-4254-4ac6-989c-c7fa0c4cc682/29468dc9-4254-4ac6-989c-c7fa0c4cc682.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1158.007310] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Created folder: Project (7da7ad544b704286a8138bfe82586c6b) in parent group-v845547. [ 1158.007580] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Creating folder: Instances. Parent ref: group-v845843. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1158.007832] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d458a2f5-a9d1-4213-b35a-7f6afeb3b9d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.009481] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-edddd267-2ce6-40fc-93aa-7e93d96ed73c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.017934] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1158.017934] env[62814]: value = "task-4294581" [ 1158.017934] env[62814]: _type = "Task" [ 1158.017934] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.025592] env[62814]: DEBUG oslo_vmware.api [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294577, 'name': PowerOffVM_Task, 'duration_secs': 0.327608} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.025873] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Created folder: Instances in parent group-v845843. [ 1158.026137] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1158.026690] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1158.026877] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1158.027099] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1158.027334] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-815f2133-9cdc-4c4e-b1db-d829f134fb40 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.028797] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cbfeaef4-00bc-4197-99cd-a62bcc3680fb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.044049] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294581, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.051358] env[62814]: DEBUG nova.compute.manager [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1158.053995] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1158.053995] env[62814]: value = "task-4294584" [ 1158.053995] env[62814]: _type = "Task" [ 1158.053995] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.062988] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294584, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.159548] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1158.160367] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1158.160367] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleting the datastore file [datastore2] 7e849b23-226d-4a1d-a85c-cf96964ce034 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1158.160367] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-237ca9f7-c93f-417f-a8ca-1be693ff8f66 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.169374] env[62814]: DEBUG oslo_vmware.api [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1158.169374] env[62814]: value = "task-4294585" [ 1158.169374] env[62814]: _type = "Task" [ 1158.169374] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.178456] env[62814]: DEBUG oslo_vmware.api [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.270040] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294574, 'name': CloneVM_Task, 'duration_secs': 1.482877} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.270351] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Created linked-clone VM from snapshot [ 1158.271414] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c87db3-9412-4bb8-aee2-aa8a8e144d8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.281928] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Uploading image d3d599ce-93b8-4dd8-a28d-2a1abc5ea448 {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1158.308289] env[62814]: DEBUG oslo_vmware.rw_handles [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1158.308289] env[62814]: value = "vm-845842" [ 1158.308289] env[62814]: _type = "VirtualMachine" [ 1158.308289] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1158.308570] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c6ff5b54-4983-4329-b338-f84fd5ab599c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.317408] env[62814]: DEBUG oslo_vmware.rw_handles [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lease: (returnval){ [ 1158.317408] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527101c7-fb51-e3d9-8490-080ddb1b93bf" [ 1158.317408] env[62814]: _type = "HttpNfcLease" [ 1158.317408] env[62814]: } obtained for exporting VM: (result){ [ 1158.317408] env[62814]: value = "vm-845842" [ 1158.317408] env[62814]: _type = "VirtualMachine" [ 1158.317408] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1158.318360] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the lease: (returnval){ [ 1158.318360] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527101c7-fb51-e3d9-8490-080ddb1b93bf" [ 1158.318360] env[62814]: _type = "HttpNfcLease" [ 1158.318360] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1158.329995] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1158.329995] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527101c7-fb51-e3d9-8490-080ddb1b93bf" [ 1158.329995] env[62814]: _type = "HttpNfcLease" [ 1158.329995] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1158.433022] env[62814]: DEBUG oslo_vmware.api [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294579, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256938} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.435707] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1158.435899] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1158.436104] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1158.436279] env[62814]: INFO nova.compute.manager [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1158.436520] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1158.437601] env[62814]: DEBUG nova.compute.manager [-] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1158.437809] env[62814]: DEBUG nova.network.neutron [-] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1158.528070] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294581, 'name': Rename_Task, 'duration_secs': 0.205352} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.530791] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1158.531345] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6d72dc6f-c49f-49ae-8e1a-941c8cef7366 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.540012] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1158.540012] env[62814]: value = "task-4294587" [ 1158.540012] env[62814]: _type = "Task" [ 1158.540012] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.552261] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294587, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.569520] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294584, 'name': CreateVM_Task, 'duration_secs': 0.309582} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.569919] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1158.570449] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1158.571443] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1158.571443] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1158.571443] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-872c631d-5737-42e6-8703-44b009e14f5a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.576702] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1158.576702] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ea82e6-61c1-b564-39d5-4a4ef014e545" [ 1158.576702] env[62814]: _type = "Task" [ 1158.576702] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.583570] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1158.589260] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ea82e6-61c1-b564-39d5-4a4ef014e545, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.607676] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b5b6a61-7b0a-4689-b88f-2acc724ec5db {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.615765] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96514d62-1590-48ba-9671-82ddfdbcccee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.659294] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c12e34-fad8-4421-a6d6-0cd8c2987cc8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.667896] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da064046-1681-406b-ae1a-16f4a8de3b7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.686295] env[62814]: DEBUG nova.compute.provider_tree [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.691176] env[62814]: DEBUG oslo_vmware.api [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154753} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.691681] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1158.691890] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1158.692091] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1158.692331] env[62814]: INFO nova.compute.manager [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1158.692588] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1158.692778] env[62814]: DEBUG nova.compute.manager [-] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1158.692872] env[62814]: DEBUG nova.network.neutron [-] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1158.829471] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1158.829471] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527101c7-fb51-e3d9-8490-080ddb1b93bf" [ 1158.829471] env[62814]: _type = "HttpNfcLease" [ 1158.829471] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1158.829614] env[62814]: DEBUG oslo_vmware.rw_handles [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1158.829614] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527101c7-fb51-e3d9-8490-080ddb1b93bf" [ 1158.829614] env[62814]: _type = "HttpNfcLease" [ 1158.829614] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1158.830870] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc0bb01f-bbac-4cb1-abe3-9e0cd3ca3cdf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.838779] env[62814]: DEBUG oslo_vmware.rw_handles [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2949-5e34-8f06-6ff7-304dce9bf322/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1158.838989] env[62814]: DEBUG oslo_vmware.rw_handles [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2949-5e34-8f06-6ff7-304dce9bf322/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1158.909816] env[62814]: DEBUG nova.compute.manager [req-2ee84f4c-8f55-4009-a51a-3df9bfdb6480 req-fcf2c07c-26b9-44f9-9574-a450351eaae5 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Received event network-vif-deleted-bb09100c-d462-4d19-9fd5-1726001e8e55 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1158.910053] env[62814]: INFO nova.compute.manager [req-2ee84f4c-8f55-4009-a51a-3df9bfdb6480 req-fcf2c07c-26b9-44f9-9574-a450351eaae5 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Neutron deleted interface bb09100c-d462-4d19-9fd5-1726001e8e55; detaching it from the instance and deleting it from the info cache [ 1158.910241] env[62814]: DEBUG nova.network.neutron [req-2ee84f4c-8f55-4009-a51a-3df9bfdb6480 req-fcf2c07c-26b9-44f9-9574-a450351eaae5 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1158.958437] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-76cf657b-6bd1-4612-b6d9-ec1ac807b561 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.051137] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294587, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.088807] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ea82e6-61c1-b564-39d5-4a4ef014e545, 'name': SearchDatastore_Task, 'duration_secs': 0.013047} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.089207] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1159.089513] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1159.090091] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.090091] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1159.090091] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1159.090358] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4825d36-fdc9-4011-9a24-b20cc0038b57 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.099490] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1159.099695] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1159.100503] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61df893d-e804-4c3b-863c-575d9938d344 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.107107] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1159.107107] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a52c84-e0ba-b8cb-8808-2cfe57bbd0ef" [ 1159.107107] env[62814]: _type = "Task" [ 1159.107107] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.119626] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a52c84-e0ba-b8cb-8808-2cfe57bbd0ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.193754] env[62814]: DEBUG nova.scheduler.client.report [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1159.303899] env[62814]: DEBUG nova.network.neutron [-] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.413278] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb687dfe-b25d-49df-8fd8-134864cf8a64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.423849] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9cefe5-5894-4f00-abb4-89c4f3b9ea75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.438408] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1159.438961] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.460288] env[62814]: DEBUG nova.compute.manager [req-2ee84f4c-8f55-4009-a51a-3df9bfdb6480 req-fcf2c07c-26b9-44f9-9574-a450351eaae5 service nova] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Detach interface failed, port_id=bb09100c-d462-4d19-9fd5-1726001e8e55, reason: Instance 201cf7be-f81f-4eba-ae65-a370ad8e48c3 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1159.484213] env[62814]: DEBUG nova.network.neutron [-] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.551568] env[62814]: DEBUG oslo_vmware.api [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294587, 'name': PowerOnVM_Task, 'duration_secs': 0.697556} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.552118] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1159.552616] env[62814]: INFO nova.compute.manager [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Took 6.39 seconds to spawn the instance on the hypervisor. [ 1159.553019] env[62814]: DEBUG nova.compute.manager [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1159.554383] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b511d634-a9e2-4cfc-8baa-a54d3594d2ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.621743] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a52c84-e0ba-b8cb-8808-2cfe57bbd0ef, 'name': SearchDatastore_Task, 'duration_secs': 0.011533} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.622716] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5071904-f19c-4ded-9cfa-a55e37165674 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.628888] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1159.628888] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52798a2c-5ba4-d03e-7a01-ebd21c17a78d" [ 1159.628888] env[62814]: _type = "Task" [ 1159.628888] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.638928] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52798a2c-5ba4-d03e-7a01-ebd21c17a78d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.702225] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.326s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1159.702538] env[62814]: DEBUG nova.compute.manager [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1159.705755] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.247s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1159.705755] env[62814]: DEBUG nova.objects.instance [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lazy-loading 'pci_requests' on Instance uuid edcdb1e5-ed9d-49e6-97e6-ea7629682547 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1159.811488] env[62814]: INFO nova.compute.manager [-] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Took 1.37 seconds to deallocate network for instance. [ 1159.943195] env[62814]: DEBUG nova.compute.utils [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1159.986883] env[62814]: INFO nova.compute.manager [-] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Took 1.29 seconds to deallocate network for instance. [ 1160.076030] env[62814]: INFO nova.compute.manager [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Took 30.32 seconds to build instance. [ 1160.140021] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52798a2c-5ba4-d03e-7a01-ebd21c17a78d, 'name': SearchDatastore_Task, 'duration_secs': 0.018288} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.140526] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1160.140915] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 53e13866-0174-4867-9fd9-2ea3f5ee399d/53e13866-0174-4867-9fd9-2ea3f5ee399d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1160.141264] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a4080148-1356-4e89-9595-e737d5c6f9e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.148866] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1160.148866] env[62814]: value = "task-4294588" [ 1160.148866] env[62814]: _type = "Task" [ 1160.148866] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.157438] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294588, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.210191] env[62814]: DEBUG nova.compute.utils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1160.213228] env[62814]: DEBUG nova.objects.instance [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lazy-loading 'numa_topology' on Instance uuid edcdb1e5-ed9d-49e6-97e6-ea7629682547 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1160.215176] env[62814]: DEBUG nova.compute.manager [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1160.215394] env[62814]: DEBUG nova.network.neutron [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1160.278026] env[62814]: DEBUG nova.policy [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e246f49b0b84cd093549b6d6b384e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6e52480dd2c467790622901940cf385', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1160.322672] env[62814]: DEBUG oslo_concurrency.lockutils [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.446338] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1160.496259] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1160.552638] env[62814]: DEBUG nova.network.neutron [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Successfully created port: 26a8d6bc-23fd-40f8-ac9c-cea10954ca94 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1160.578521] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c05c2a10-3b3a-4561-b50b-08fa80bbdb5b tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "29468dc9-4254-4ac6-989c-c7fa0c4cc682" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.833s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1160.659887] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294588, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.716377] env[62814]: DEBUG nova.compute.manager [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1160.721237] env[62814]: INFO nova.compute.claims [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1160.730031] env[62814]: INFO nova.compute.manager [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Rebuilding instance [ 1160.790450] env[62814]: DEBUG nova.compute.manager [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1160.791432] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea6fc47-fe82-4b72-aec6-09b1dc0ec81f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.936437] env[62814]: DEBUG nova.compute.manager [req-dac8550f-a746-4d36-acdd-ef2fa7438bd6 req-cecd8842-bec9-4a89-90d5-8a690669d0c4 service nova] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Received event network-vif-deleted-f1881a6c-5678-4b14-9eb9-f3e7f3e41810 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1161.160284] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294588, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.733874} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.160729] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 53e13866-0174-4867-9fd9-2ea3f5ee399d/53e13866-0174-4867-9fd9-2ea3f5ee399d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1161.161053] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1161.161385] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59ea1590-31c2-43b1-9a87-64b90a81a818 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.168511] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1161.168511] env[62814]: value = "task-4294589" [ 1161.168511] env[62814]: _type = "Task" [ 1161.168511] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.177870] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294589, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.430123] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b364e6-653f-4464-b8c4-b4f3fc9c1d79 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.438507] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b41872-0275-4d6e-b609-6c1f17cfd8e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.472340] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc2be73-6b02-44b7-a3f5-cce7b4a69969 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.480923] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e088a2-874e-493a-85d1-03c3af4268c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.496158] env[62814]: DEBUG nova.compute.provider_tree [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1161.503080] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1161.503520] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1161.503754] env[62814]: INFO nova.compute.manager [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Attaching volume 13a381a6-dd99-479c-a496-196303cb75b3 to /dev/sdb [ 1161.538087] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffbf4800-9449-45c9-b767-9e8615115d84 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.546029] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d02720b-b045-4ca1-8010-9c92c70f4859 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.563099] env[62814]: DEBUG nova.virt.block_device [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Updating existing volume attachment record: 5cffe135-5c59-4727-8231-e464032e7acf {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1161.679430] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294589, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067035} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1161.679736] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1161.680593] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39efb854-7ae2-412d-a578-9af18aa6c4ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.701211] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 53e13866-0174-4867-9fd9-2ea3f5ee399d/53e13866-0174-4867-9fd9-2ea3f5ee399d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1161.701560] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89402800-1280-453a-8aae-aa074f7a8052 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.721480] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1161.721480] env[62814]: value = "task-4294590" [ 1161.721480] env[62814]: _type = "Task" [ 1161.721480] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.730200] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294590, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.731531] env[62814]: DEBUG nova.compute.manager [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1161.760758] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1161.761076] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1161.761276] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1161.761539] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1161.761757] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1161.761757] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1161.761970] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1161.762148] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1161.762456] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1161.762636] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1161.762831] env[62814]: DEBUG nova.virt.hardware [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1161.763784] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf4a85f-1f43-43ab-aee2-4a00382bd1a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.774138] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bbaa09-075e-489b-addd-229a37884eee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.805459] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1161.805790] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5e0e679-a6dd-4b95-83c2-4c9f2e140e7d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.812907] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1161.812907] env[62814]: value = "task-4294592" [ 1161.812907] env[62814]: _type = "Task" [ 1161.812907] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1161.822272] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294592, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.999882] env[62814]: DEBUG nova.scheduler.client.report [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1162.186261] env[62814]: DEBUG nova.network.neutron [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Successfully updated port: 26a8d6bc-23fd-40f8-ac9c-cea10954ca94 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1162.235134] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294590, 'name': ReconfigVM_Task, 'duration_secs': 0.387744} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.235429] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 53e13866-0174-4867-9fd9-2ea3f5ee399d/53e13866-0174-4867-9fd9-2ea3f5ee399d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1162.236070] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f2e2d867-4801-4b10-a498-88ad1569c9e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.244408] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1162.244408] env[62814]: value = "task-4294593" [ 1162.244408] env[62814]: _type = "Task" [ 1162.244408] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.255992] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294593, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.322794] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294592, 'name': PowerOffVM_Task, 'duration_secs': 0.163787} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.323094] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1162.323375] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1162.324211] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36294fc6-031e-486f-afe7-f1f338875b3c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.331518] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1162.332089] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c632567f-b976-447f-ba8f-376125cdcbdf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.358467] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1162.358800] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1162.359050] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Deleting the datastore file [datastore2] 29468dc9-4254-4ac6-989c-c7fa0c4cc682 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1162.359376] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-06682fb8-6a21-43ff-9afe-95cfac919803 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.366271] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1162.366271] env[62814]: value = "task-4294595" [ 1162.366271] env[62814]: _type = "Task" [ 1162.366271] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.375378] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294595, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.507042] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.801s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1162.508734] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.925s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.510295] env[62814]: INFO nova.compute.claims [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1162.543475] env[62814]: INFO nova.network.neutron [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updating port 9950de47-d55a-480b-b88e-818e09b9bc80 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1162.689482] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "refresh_cache-881e9481-a5a9-489b-8abd-8efdaa239a11" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1162.689482] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "refresh_cache-881e9481-a5a9-489b-8abd-8efdaa239a11" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1162.689482] env[62814]: DEBUG nova.network.neutron [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1162.755780] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294593, 'name': Rename_Task, 'duration_secs': 0.144649} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.756149] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1162.756433] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-30640c25-18ba-40ab-88bf-89f54ba58944 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.763126] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1162.763126] env[62814]: value = "task-4294596" [ 1162.763126] env[62814]: _type = "Task" [ 1162.763126] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.770985] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294596, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.876317] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294595, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151909} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.876571] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1162.876756] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1162.876933] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1162.965983] env[62814]: DEBUG nova.compute.manager [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Received event network-vif-plugged-26a8d6bc-23fd-40f8-ac9c-cea10954ca94 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1162.966563] env[62814]: DEBUG oslo_concurrency.lockutils [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] Acquiring lock "881e9481-a5a9-489b-8abd-8efdaa239a11-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1162.966895] env[62814]: DEBUG oslo_concurrency.lockutils [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] Lock "881e9481-a5a9-489b-8abd-8efdaa239a11-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1162.967167] env[62814]: DEBUG oslo_concurrency.lockutils [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] Lock "881e9481-a5a9-489b-8abd-8efdaa239a11-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1162.967453] env[62814]: DEBUG nova.compute.manager [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] No waiting events found dispatching network-vif-plugged-26a8d6bc-23fd-40f8-ac9c-cea10954ca94 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1162.967700] env[62814]: WARNING nova.compute.manager [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Received unexpected event network-vif-plugged-26a8d6bc-23fd-40f8-ac9c-cea10954ca94 for instance with vm_state building and task_state spawning. [ 1162.967934] env[62814]: DEBUG nova.compute.manager [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Received event network-changed-26a8d6bc-23fd-40f8-ac9c-cea10954ca94 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1162.968227] env[62814]: DEBUG nova.compute.manager [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Refreshing instance network info cache due to event network-changed-26a8d6bc-23fd-40f8-ac9c-cea10954ca94. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1162.968484] env[62814]: DEBUG oslo_concurrency.lockutils [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] Acquiring lock "refresh_cache-881e9481-a5a9-489b-8abd-8efdaa239a11" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1163.225334] env[62814]: DEBUG nova.network.neutron [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1163.275754] env[62814]: DEBUG oslo_vmware.api [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294596, 'name': PowerOnVM_Task, 'duration_secs': 0.511299} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1163.276147] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1163.276435] env[62814]: INFO nova.compute.manager [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Took 5.35 seconds to spawn the instance on the hypervisor. [ 1163.276679] env[62814]: DEBUG nova.compute.manager [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1163.277547] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a45a0ae-8cfe-4724-9032-a2636a62cfde {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.374734] env[62814]: DEBUG nova.network.neutron [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Updating instance_info_cache with network_info: [{"id": "26a8d6bc-23fd-40f8-ac9c-cea10954ca94", "address": "fa:16:3e:25:d9:b4", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26a8d6bc-23", "ovs_interfaceid": "26a8d6bc-23fd-40f8-ac9c-cea10954ca94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1163.723430] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5fbc790-2fa6-4e8c-a61b-3a7d11083a79 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.731682] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743be52c-1fc3-442b-91d4-1b339ff8db47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.764336] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617da958-a59a-4a05-b7b1-fef7bac182bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.771735] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e0987f-715f-4ed7-966a-8167e480a61e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.786580] env[62814]: DEBUG nova.compute.provider_tree [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1163.799370] env[62814]: INFO nova.compute.manager [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Took 23.73 seconds to build instance. [ 1163.878069] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "refresh_cache-881e9481-a5a9-489b-8abd-8efdaa239a11" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1163.878069] env[62814]: DEBUG nova.compute.manager [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Instance network_info: |[{"id": "26a8d6bc-23fd-40f8-ac9c-cea10954ca94", "address": "fa:16:3e:25:d9:b4", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26a8d6bc-23", "ovs_interfaceid": "26a8d6bc-23fd-40f8-ac9c-cea10954ca94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1163.878377] env[62814]: DEBUG oslo_concurrency.lockutils [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] Acquired lock "refresh_cache-881e9481-a5a9-489b-8abd-8efdaa239a11" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1163.878517] env[62814]: DEBUG nova.network.neutron [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Refreshing network info cache for port 26a8d6bc-23fd-40f8-ac9c-cea10954ca94 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1163.879850] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:25:d9:b4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '51bac3c3-00ab-4a07-9e28-b3c951dee565', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '26a8d6bc-23fd-40f8-ac9c-cea10954ca94', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1163.887873] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1163.891536] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1163.892271] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8f97e1c-2126-4394-adfa-f759b06ee4e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.923530] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1163.923784] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1163.923937] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1163.924160] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1163.924798] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1163.924798] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1163.924928] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1163.925165] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1163.925476] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1163.925656] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1163.925829] env[62814]: DEBUG nova.virt.hardware [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1163.927802] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7644469b-a71d-46b4-8cd5-c00fb282da34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.930704] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1163.930704] env[62814]: value = "task-4294598" [ 1163.930704] env[62814]: _type = "Task" [ 1163.930704] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.938118] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745a0eb5-ffaa-46c4-b62a-9c05927fdfb2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.945753] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294598, 'name': CreateVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1163.957996] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1163.963744] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1163.964040] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1163.964283] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6e5b6a3-1dcb-4128-97a5-19b72ca4dd0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.982129] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1163.982129] env[62814]: value = "task-4294599" [ 1163.982129] env[62814]: _type = "Task" [ 1163.982129] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1163.990514] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294599, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.084359] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.084553] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.084784] env[62814]: DEBUG nova.network.neutron [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1164.292123] env[62814]: DEBUG nova.scheduler.client.report [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1164.301774] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a30d5ff5-0683-41fa-82f2-e01ec036cd89 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "53e13866-0174-4867-9fd9-2ea3f5ee399d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.240s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.337409] env[62814]: INFO nova.compute.manager [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Rebuilding instance [ 1164.379913] env[62814]: DEBUG nova.compute.manager [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1164.380808] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d703fba-77fd-4425-bf39-bd6f9c9a100b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.440247] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294598, 'name': CreateVM_Task, 'duration_secs': 0.410517} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.440417] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1164.441077] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.441276] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.441610] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1164.441864] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6060751-c6d6-4e3e-8e39-e66633691922 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.446952] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1164.446952] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d57b65-7005-188b-a8b5-96f7eefbea5b" [ 1164.446952] env[62814]: _type = "Task" [ 1164.446952] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.454684] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d57b65-7005-188b-a8b5-96f7eefbea5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.492663] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294599, 'name': CreateVM_Task, 'duration_secs': 0.36251} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.492847] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1164.493285] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.506428] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1164.506665] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1164.664022] env[62814]: DEBUG nova.network.neutron [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Updated VIF entry in instance network info cache for port 26a8d6bc-23fd-40f8-ac9c-cea10954ca94. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1164.664022] env[62814]: DEBUG nova.network.neutron [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Updating instance_info_cache with network_info: [{"id": "26a8d6bc-23fd-40f8-ac9c-cea10954ca94", "address": "fa:16:3e:25:d9:b4", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap26a8d6bc-23", "ovs_interfaceid": "26a8d6bc-23fd-40f8-ac9c-cea10954ca94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.797642] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.289s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.798185] env[62814]: DEBUG nova.compute.manager [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1164.800806] env[62814]: DEBUG oslo_concurrency.lockutils [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.479s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.801029] env[62814]: DEBUG nova.objects.instance [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lazy-loading 'resources' on Instance uuid 201cf7be-f81f-4eba-ae65-a370ad8e48c3 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.825095] env[62814]: DEBUG nova.network.neutron [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updating instance_info_cache with network_info: [{"id": "9950de47-d55a-480b-b88e-818e09b9bc80", "address": "fa:16:3e:0c:71:2f", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9950de47-d5", "ovs_interfaceid": "9950de47-d55a-480b-b88e-818e09b9bc80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1164.959069] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d57b65-7005-188b-a8b5-96f7eefbea5b, 'name': SearchDatastore_Task, 'duration_secs': 0.013356} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1164.959399] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1164.959639] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1164.959881] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.960040] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.960227] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1164.960515] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1164.960900] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1164.961161] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bce623b3-ab55-4cbe-a9c6-730313dd3fa1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.963027] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb33de80-3d67-4eca-9138-564ba1fe8d84 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.968763] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1164.968763] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525caa43-83b9-3a77-be19-22b24ee05aaa" [ 1164.968763] env[62814]: _type = "Task" [ 1164.968763] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.973278] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1164.973479] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1164.974509] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ecd65b4e-0023-41c0-8b52-439d570f9079 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.979726] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525caa43-83b9-3a77-be19-22b24ee05aaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.982647] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1164.982647] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522e1edd-bbe6-0193-20e3-8290c07891da" [ 1164.982647] env[62814]: _type = "Task" [ 1164.982647] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.989915] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522e1edd-bbe6-0193-20e3-8290c07891da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1164.996933] env[62814]: DEBUG nova.compute.manager [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Received event network-vif-plugged-9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1164.997163] env[62814]: DEBUG oslo_concurrency.lockutils [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] Acquiring lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1164.997369] env[62814]: DEBUG oslo_concurrency.lockutils [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1164.997552] env[62814]: DEBUG oslo_concurrency.lockutils [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1164.997705] env[62814]: DEBUG nova.compute.manager [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] No waiting events found dispatching network-vif-plugged-9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1164.997869] env[62814]: WARNING nova.compute.manager [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Received unexpected event network-vif-plugged-9950de47-d55a-480b-b88e-818e09b9bc80 for instance with vm_state shelved_offloaded and task_state spawning. [ 1164.998085] env[62814]: DEBUG nova.compute.manager [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Received event network-changed-9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1164.998199] env[62814]: DEBUG nova.compute.manager [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Refreshing instance network info cache due to event network-changed-9950de47-d55a-480b-b88e-818e09b9bc80. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1164.998339] env[62814]: DEBUG oslo_concurrency.lockutils [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] Acquiring lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.009538] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1165.166181] env[62814]: DEBUG oslo_concurrency.lockutils [req-7535a10a-b35b-4cf7-92bc-97ee9d9daf40 req-d125d7a8-c92d-4380-8c75-3832d119728c service nova] Releasing lock "refresh_cache-881e9481-a5a9-489b-8abd-8efdaa239a11" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1165.305053] env[62814]: DEBUG nova.compute.utils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1165.309171] env[62814]: DEBUG nova.compute.manager [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1165.309388] env[62814]: DEBUG nova.network.neutron [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1165.328035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1165.332654] env[62814]: DEBUG oslo_concurrency.lockutils [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] Acquired lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1165.332848] env[62814]: DEBUG nova.network.neutron [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Refreshing network info cache for port 9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1165.359833] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='a7349042b4362f5d7e8a6c79bbecb10e',container_format='bare',created_at=2025-06-21T05:42:46Z,direct_url=,disk_format='vmdk',id=dc3cbb42-9340-472a-b42a-4fb9bea7421b,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-2146256746-shelved',owner='869db97f30f544f49e6a0c0e54bea463',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2025-06-21T05:43:04Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1165.360122] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1165.360304] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1165.360517] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1165.360666] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1165.360832] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1165.361085] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1165.361295] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1165.361508] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1165.361679] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1165.361853] env[62814]: DEBUG nova.virt.hardware [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1165.363067] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd48755a-3720-4bbc-804a-6290ae638ae8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.373724] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae174b9-3f9b-4528-bb5e-c20b66df8844 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.390695] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:71:2f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ad4fcde7-8926-402a-a9b7-4878d2bc1cf6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9950de47-d55a-480b-b88e-818e09b9bc80', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1165.398450] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1165.400113] env[62814]: DEBUG nova.policy [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64e84da705284d2dbf693a26ef184cd7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95c92336f9e746edba50b0b9e078b0dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1165.404253] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1165.404522] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1165.404962] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eb299b99-9a7e-4019-b420-2f3e4aa42fb8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.406566] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4d621276-7397-41cb-8074-ce64dce235b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.430811] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1165.430811] env[62814]: value = "task-4294600" [ 1165.430811] env[62814]: _type = "Task" [ 1165.430811] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.431890] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1165.431890] env[62814]: value = "task-4294601" [ 1165.431890] env[62814]: _type = "Task" [ 1165.431890] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.447616] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.451151] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294601, 'name': CreateVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.482494] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525caa43-83b9-3a77-be19-22b24ee05aaa, 'name': SearchDatastore_Task, 'duration_secs': 0.014506} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.482845] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1165.483101] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1165.483377] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.498247] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522e1edd-bbe6-0193-20e3-8290c07891da, 'name': SearchDatastore_Task, 'duration_secs': 0.013056} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.499212] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27b27a1b-0efe-455c-8db2-823da96ed79b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.509361] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1165.509361] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d458be-ef17-ab6a-fec7-15bbaa5bf6ff" [ 1165.509361] env[62814]: _type = "Task" [ 1165.509361] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.518177] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d458be-ef17-ab6a-fec7-15bbaa5bf6ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.587614] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403507bc-0b1a-49e1-8970-e07b62cdb4cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.596626] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a091b07d-bf56-40e5-9b1f-f9a0b6941d09 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.627755] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86bf3fbb-b25f-4261-81f7-8f6cf38bcd34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.635863] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d584e38-17be-4443-ab62-52deaf0db6b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.650780] env[62814]: DEBUG nova.compute.provider_tree [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1165.809228] env[62814]: DEBUG nova.network.neutron [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Successfully created port: 580abcb7-b7f2-437a-8a84-738e83af0751 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1165.815188] env[62814]: DEBUG nova.compute.manager [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1165.945034] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294600, 'name': PowerOffVM_Task, 'duration_secs': 0.151084} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.948543] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1165.948964] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1165.949151] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294601, 'name': CreateVM_Task, 'duration_secs': 0.431427} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.949841] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddd4381-da24-477e-94e9-6f6447444fbc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.952590] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1165.953467] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1165.953630] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1165.954024] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1165.954659] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00d6978b-9aa6-4b71-83dd-a58cbf18dc84 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.960513] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1165.961756] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ecce5cf-7802-4a9e-9ab4-c379a82ebcf9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.966242] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1165.966242] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bc2e6e-ff57-fdf0-d918-10fc6f71b03e" [ 1165.966242] env[62814]: _type = "Task" [ 1165.966242] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.971177] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bc2e6e-ff57-fdf0-d918-10fc6f71b03e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.987374] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1165.987595] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1165.987780] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Deleting the datastore file [datastore2] 53e13866-0174-4867-9fd9-2ea3f5ee399d {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1165.988062] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-452f6902-35c2-458c-958a-71fd4ad1d8f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.994869] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1165.994869] env[62814]: value = "task-4294603" [ 1165.994869] env[62814]: _type = "Task" [ 1165.994869] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.003322] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294603, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.018636] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d458be-ef17-ab6a-fec7-15bbaa5bf6ff, 'name': SearchDatastore_Task, 'duration_secs': 0.016787} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.018898] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1166.019177] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 881e9481-a5a9-489b-8abd-8efdaa239a11/881e9481-a5a9-489b-8abd-8efdaa239a11.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1166.019487] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1166.019674] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1166.019891] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc520a44-f1e1-4425-866d-8392af75f134 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.021882] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef02ca41-8778-4460-9754-f45c14968997 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.029032] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1166.029032] env[62814]: value = "task-4294604" [ 1166.029032] env[62814]: _type = "Task" [ 1166.029032] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.036996] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294604, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.039267] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1166.039467] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1166.042942] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d1ab673-854e-4993-b21c-962860de96d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.048716] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1166.048716] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d98fb5-fcbc-3c2d-f064-7c504463e7e1" [ 1166.048716] env[62814]: _type = "Task" [ 1166.048716] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.056780] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d98fb5-fcbc-3c2d-f064-7c504463e7e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.115256] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1166.115501] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845846', 'volume_id': '13a381a6-dd99-479c-a496-196303cb75b3', 'name': 'volume-13a381a6-dd99-479c-a496-196303cb75b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f00d414e-1dfd-43cf-9245-7cbeea8850b3', 'attached_at': '', 'detached_at': '', 'volume_id': '13a381a6-dd99-479c-a496-196303cb75b3', 'serial': '13a381a6-dd99-479c-a496-196303cb75b3'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1166.116435] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5809b522-60ce-4148-a4b6-65cc04d19385 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.133324] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4b0f55-83fe-4e6e-8086-61a7cc46016f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.158167] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] volume-13a381a6-dd99-479c-a496-196303cb75b3/volume-13a381a6-dd99-479c-a496-196303cb75b3.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1166.161753] env[62814]: DEBUG nova.scheduler.client.report [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1166.165034] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a75d5910-55f9-40ad-96ae-ba0edf96adae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.183695] env[62814]: DEBUG oslo_concurrency.lockutils [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.383s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.186162] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.690s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.186401] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.188087] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.179s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.188276] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.188453] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1166.190793] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50dc177d-b159-4cf1-877b-25f44ca993ac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.196663] env[62814]: DEBUG oslo_vmware.api [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1166.196663] env[62814]: value = "task-4294605" [ 1166.196663] env[62814]: _type = "Task" [ 1166.196663] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.204357] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8704b21f-4873-41ec-8517-1b3007fd11e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.225796] env[62814]: DEBUG oslo_vmware.api [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294605, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.227585] env[62814]: INFO nova.scheduler.client.report [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted allocations for instance 7e849b23-226d-4a1d-a85c-cf96964ce034 [ 1166.231484] env[62814]: INFO nova.scheduler.client.report [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted allocations for instance 201cf7be-f81f-4eba-ae65-a370ad8e48c3 [ 1166.232740] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb420e1e-5b56-4247-8d6f-9ede088a6ede {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.243511] env[62814]: DEBUG nova.network.neutron [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updated VIF entry in instance network info cache for port 9950de47-d55a-480b-b88e-818e09b9bc80. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1166.243949] env[62814]: DEBUG nova.network.neutron [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updating instance_info_cache with network_info: [{"id": "9950de47-d55a-480b-b88e-818e09b9bc80", "address": "fa:16:3e:0c:71:2f", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9950de47-d5", "ovs_interfaceid": "9950de47-d55a-480b-b88e-818e09b9bc80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.250640] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ded81e0b-a960-4cf9-a9b2-e43fd9140519 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.288049] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179598MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1166.288049] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1166.288049] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1166.476255] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1166.476537] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Processing image dc3cbb42-9340-472a-b42a-4fb9bea7421b {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1166.476786] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b/dc3cbb42-9340-472a-b42a-4fb9bea7421b.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1166.476937] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "[datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b/dc3cbb42-9340-472a-b42a-4fb9bea7421b.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1166.477122] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1166.477428] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5f37a00d-8088-4a56-966f-ea3df53f02d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.492394] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1166.492749] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1166.494027] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1200e7d-2dc7-4be0-80fb-1ba53fa1d395 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.506137] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1166.506137] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529e0599-2c7b-f488-b795-38020ced5519" [ 1166.506137] env[62814]: _type = "Task" [ 1166.506137] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.510953] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294603, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163424} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.514494] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1166.514727] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1166.514962] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1166.523080] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529e0599-2c7b-f488-b795-38020ced5519, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.541356] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294604, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.563078] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d98fb5-fcbc-3c2d-f064-7c504463e7e1, 'name': SearchDatastore_Task, 'duration_secs': 0.01385} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.565376] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8848d44b-b557-4c26-ac34-d2026809a7d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.569943] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1166.569943] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528d0905-194b-1d9c-bc8b-77a774293224" [ 1166.569943] env[62814]: _type = "Task" [ 1166.569943] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.578450] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528d0905-194b-1d9c-bc8b-77a774293224, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.708716] env[62814]: DEBUG oslo_vmware.api [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294605, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.745692] env[62814]: DEBUG oslo_concurrency.lockutils [None req-68459a82-2e34-4d6b-a4af-f0f34aac1408 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "201cf7be-f81f-4eba-ae65-a370ad8e48c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.926s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.746408] env[62814]: DEBUG oslo_concurrency.lockutils [req-735099dd-7c23-4ff0-aebb-1cb5be2eb05a req-effb8d06-99c1-49ae-b454-5ee029aed604 service nova] Releasing lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1166.747074] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29264b3-14d5-48ed-a781-3c15d88ad91d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "7e849b23-226d-4a1d-a85c-cf96964ce034" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.813s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1166.825099] env[62814]: DEBUG nova.compute.manager [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1166.855422] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1166.855775] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1166.856078] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1166.856301] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1166.856567] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1166.856778] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1166.857009] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1166.857290] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1166.857534] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1166.857711] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1166.858063] env[62814]: DEBUG nova.virt.hardware [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1166.859196] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6efdb53-ba54-4c82-939f-1aa52c4e1c4a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.868509] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd43bb8a-5f61-4bd2-9253-9253fc60eb6f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.966861] env[62814]: DEBUG oslo_vmware.rw_handles [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2949-5e34-8f06-6ff7-304dce9bf322/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1166.967836] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26bb9b2-60ac-422a-b15e-54c68a1d70e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.973936] env[62814]: DEBUG oslo_vmware.rw_handles [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2949-5e34-8f06-6ff7-304dce9bf322/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1166.974089] env[62814]: ERROR oslo_vmware.rw_handles [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2949-5e34-8f06-6ff7-304dce9bf322/disk-0.vmdk due to incomplete transfer. [ 1166.974530] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f68d66cf-fb14-41fb-b721-dfe230676b01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.981540] env[62814]: DEBUG oslo_vmware.rw_handles [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2949-5e34-8f06-6ff7-304dce9bf322/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1166.981823] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Uploaded image d3d599ce-93b8-4dd8-a28d-2a1abc5ea448 to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1166.983949] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1166.984201] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-813b25fd-140f-4ce1-b5de-35bd37484595 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.992014] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1166.992014] env[62814]: value = "task-4294606" [ 1166.992014] env[62814]: _type = "Task" [ 1166.992014] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.998549] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294606, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.023649] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Preparing fetch location {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1167.023818] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Fetch image to [datastore2] OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2/OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2.vmdk {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1167.024057] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Downloading stream optimized image dc3cbb42-9340-472a-b42a-4fb9bea7421b to [datastore2] OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2/OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2.vmdk on the data store datastore2 as vApp {{(pid=62814) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1167.024215] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Downloading image file data dc3cbb42-9340-472a-b42a-4fb9bea7421b to the ESX as VM named 'OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2' {{(pid=62814) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1167.039428] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294604, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575893} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.039699] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 881e9481-a5a9-489b-8abd-8efdaa239a11/881e9481-a5a9-489b-8abd-8efdaa239a11.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1167.039907] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1167.040226] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b6fc502-5695-4c93-af07-d328c6b4b85c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.048889] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1167.048889] env[62814]: value = "task-4294607" [ 1167.048889] env[62814]: _type = "Task" [ 1167.048889] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.080858] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294607, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.085793] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528d0905-194b-1d9c-bc8b-77a774293224, 'name': SearchDatastore_Task, 'duration_secs': 0.019985} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.086281] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1167.086628] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 29468dc9-4254-4ac6-989c-c7fa0c4cc682/29468dc9-4254-4ac6-989c-c7fa0c4cc682.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1167.086889] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-141479eb-7fba-41cc-ac15-53cd1602bc9d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.093471] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1167.093471] env[62814]: value = "task-4294608" [ 1167.093471] env[62814]: _type = "Task" [ 1167.093471] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.103472] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294608, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.107508] env[62814]: DEBUG oslo_vmware.rw_handles [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1167.107508] env[62814]: value = "resgroup-9" [ 1167.107508] env[62814]: _type = "ResourcePool" [ 1167.107508] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1167.107807] env[62814]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d113fcbf-2fc3-464e-aba7-80b570c458f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.127840] env[62814]: DEBUG oslo_vmware.rw_handles [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lease: (returnval){ [ 1167.127840] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209cea5-dea1-7aee-8253-3c6c045059a0" [ 1167.127840] env[62814]: _type = "HttpNfcLease" [ 1167.127840] env[62814]: } obtained for vApp import into resource pool (val){ [ 1167.127840] env[62814]: value = "resgroup-9" [ 1167.127840] env[62814]: _type = "ResourcePool" [ 1167.127840] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1167.128167] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the lease: (returnval){ [ 1167.128167] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209cea5-dea1-7aee-8253-3c6c045059a0" [ 1167.128167] env[62814]: _type = "HttpNfcLease" [ 1167.128167] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1167.134856] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1167.134856] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209cea5-dea1-7aee-8253-3c6c045059a0" [ 1167.134856] env[62814]: _type = "HttpNfcLease" [ 1167.134856] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1167.208549] env[62814]: DEBUG oslo_vmware.api [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294605, 'name': ReconfigVM_Task, 'duration_secs': 0.647089} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.208850] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Reconfigured VM instance instance-00000063 to attach disk [datastore1] volume-13a381a6-dd99-479c-a496-196303cb75b3/volume-13a381a6-dd99-479c-a496-196303cb75b3.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1167.213756] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-697e4479-7658-483e-bc6a-fde6defdd875 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.230009] env[62814]: DEBUG oslo_vmware.api [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1167.230009] env[62814]: value = "task-4294610" [ 1167.230009] env[62814]: _type = "Task" [ 1167.230009] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.239433] env[62814]: DEBUG oslo_vmware.api [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294610, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.319884] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 7350d352-9336-40b8-81a6-0a4795d9f8dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.320040] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 1f13051e-e51b-4981-9445-d5420c3c1818 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.320173] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance f00d414e-1dfd-43cf-9245-7cbeea8850b3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.320291] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 2d7a18f3-456b-470c-a759-632e5d79a8f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.320407] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance b1d88997-e52f-41bd-b1b4-dd096d20d60a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.320519] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 33ab2d9e-e960-40ae-acf9-4fea0d9c830e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.320632] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 606e31eb-2349-427f-9c9b-ed9dc5b385f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.320742] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d65ccde2-b31a-4032-8795-1d609fdfcc73 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.320853] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 29468dc9-4254-4ac6-989c-c7fa0c4cc682 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.320960] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 53e13866-0174-4867-9fd9-2ea3f5ee399d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.321084] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance edcdb1e5-ed9d-49e6-97e6-ea7629682547 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.321204] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 881e9481-a5a9-489b-8abd-8efdaa239a11 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.321412] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 40e3c00b-2129-476e-bc67-a0ef13781990 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1167.427454] env[62814]: DEBUG nova.compute.manager [req-5ff07b0d-a5ef-4708-8ac7-912f2dd1fbcf req-913def75-78ff-4f4d-b02a-73abe2d25c6e service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Received event network-vif-plugged-580abcb7-b7f2-437a-8a84-738e83af0751 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1167.427698] env[62814]: DEBUG oslo_concurrency.lockutils [req-5ff07b0d-a5ef-4708-8ac7-912f2dd1fbcf req-913def75-78ff-4f4d-b02a-73abe2d25c6e service nova] Acquiring lock "40e3c00b-2129-476e-bc67-a0ef13781990-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1167.427875] env[62814]: DEBUG oslo_concurrency.lockutils [req-5ff07b0d-a5ef-4708-8ac7-912f2dd1fbcf req-913def75-78ff-4f4d-b02a-73abe2d25c6e service nova] Lock "40e3c00b-2129-476e-bc67-a0ef13781990-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.428152] env[62814]: DEBUG oslo_concurrency.lockutils [req-5ff07b0d-a5ef-4708-8ac7-912f2dd1fbcf req-913def75-78ff-4f4d-b02a-73abe2d25c6e service nova] Lock "40e3c00b-2129-476e-bc67-a0ef13781990-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1167.428350] env[62814]: DEBUG nova.compute.manager [req-5ff07b0d-a5ef-4708-8ac7-912f2dd1fbcf req-913def75-78ff-4f4d-b02a-73abe2d25c6e service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] No waiting events found dispatching network-vif-plugged-580abcb7-b7f2-437a-8a84-738e83af0751 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1167.428525] env[62814]: WARNING nova.compute.manager [req-5ff07b0d-a5ef-4708-8ac7-912f2dd1fbcf req-913def75-78ff-4f4d-b02a-73abe2d25c6e service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Received unexpected event network-vif-plugged-580abcb7-b7f2-437a-8a84-738e83af0751 for instance with vm_state building and task_state spawning. [ 1167.502015] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294606, 'name': Destroy_Task, 'duration_secs': 0.316418} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.502365] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Destroyed the VM [ 1167.502652] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1167.502952] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-12eb6f14-3ab8-49a8-9b17-69c66f848033 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.510029] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1167.510029] env[62814]: value = "task-4294611" [ 1167.510029] env[62814]: _type = "Task" [ 1167.510029] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.519266] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294611, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.554856] env[62814]: DEBUG nova.network.neutron [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Successfully updated port: 580abcb7-b7f2-437a-8a84-738e83af0751 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1167.563538] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1167.563868] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1167.564429] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1167.564673] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1167.564807] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1167.565141] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1167.565890] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1167.565890] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1167.566045] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1167.566153] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1167.566325] env[62814]: DEBUG nova.virt.hardware [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1167.567281] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "bf2ba498-7c47-4985-a84a-475037deec55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1167.567510] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "bf2ba498-7c47-4985-a84a-475037deec55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1167.569345] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83daf75e-9d71-459f-a60c-997060b3c192 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.594498] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a8bcb4-cf99-450e-a0aa-61969a4914c7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.599475] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294607, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067515} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.599847] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1167.604707] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240f7a8c-2ff2-41f2-b4ee-63f4eca994d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.616157] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Instance VIF info [] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1167.622651] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1167.623954] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1167.624196] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6527850c-7774-4574-bf4f-deccc9c0aa11 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.659866] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 881e9481-a5a9-489b-8abd-8efdaa239a11/881e9481-a5a9-489b-8abd-8efdaa239a11.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1167.660118] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294608, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.503625} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.663655] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85768776-a72a-4c2b-b80b-1796891ba7ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.679679] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 29468dc9-4254-4ac6-989c-c7fa0c4cc682/29468dc9-4254-4ac6-989c-c7fa0c4cc682.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1167.680011] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1167.681768] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-519f55ad-25f0-447d-ae99-4557b8b06fcc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.687025] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1167.687025] env[62814]: value = "task-4294612" [ 1167.687025] env[62814]: _type = "Task" [ 1167.687025] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.688966] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1167.688966] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209cea5-dea1-7aee-8253-3c6c045059a0" [ 1167.688966] env[62814]: _type = "HttpNfcLease" [ 1167.688966] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1167.693666] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1167.693666] env[62814]: value = "task-4294614" [ 1167.693666] env[62814]: _type = "Task" [ 1167.693666] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.693666] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1167.693666] env[62814]: value = "task-4294613" [ 1167.693666] env[62814]: _type = "Task" [ 1167.693666] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.700510] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294612, 'name': CreateVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.706300] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294614, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.710223] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294613, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.739751] env[62814]: DEBUG oslo_vmware.api [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294610, 'name': ReconfigVM_Task, 'duration_secs': 0.175711} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.740100] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845846', 'volume_id': '13a381a6-dd99-479c-a496-196303cb75b3', 'name': 'volume-13a381a6-dd99-479c-a496-196303cb75b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f00d414e-1dfd-43cf-9245-7cbeea8850b3', 'attached_at': '', 'detached_at': '', 'volume_id': '13a381a6-dd99-479c-a496-196303cb75b3', 'serial': '13a381a6-dd99-479c-a496-196303cb75b3'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1167.824334] env[62814]: INFO nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance bf2ba498-7c47-4985-a84a-475037deec55 has allocations against this compute host but is not found in the database. [ 1167.824572] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1167.824727] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3008MB phys_disk=149GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1168.021648] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294611, 'name': RemoveSnapshot_Task, 'duration_secs': 0.436956} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.021960] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1168.022315] env[62814]: DEBUG nova.compute.manager [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1168.023203] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ada6699-9e32-49e7-9a5f-713c3ba0881e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.028778] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb2d937-1a3f-4163-9af6-4e8f7e4c2130 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.039112] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bf6346-19e3-470d-8244-1d9d5fab7112 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.071463] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.071623] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1168.071776] env[62814]: DEBUG nova.network.neutron [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1168.074278] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f682f591-fb6a-4c29-b0eb-6ba9643b85fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.077131] env[62814]: DEBUG nova.compute.manager [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1168.085792] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e323ee-894d-496b-b508-7621223ee455 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.100434] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1168.158392] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1168.158392] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209cea5-dea1-7aee-8253-3c6c045059a0" [ 1168.158392] env[62814]: _type = "HttpNfcLease" [ 1168.158392] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1168.158687] env[62814]: DEBUG oslo_vmware.rw_handles [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1168.158687] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5209cea5-dea1-7aee-8253-3c6c045059a0" [ 1168.158687] env[62814]: _type = "HttpNfcLease" [ 1168.158687] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1168.159481] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8ff6de-b225-4f53-9cc9-21ef7443c5ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.167555] env[62814]: DEBUG oslo_vmware.rw_handles [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c367dc-ed09-93ee-d6ba-2b27b29c597c/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1168.167733] env[62814]: DEBUG oslo_vmware.rw_handles [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c367dc-ed09-93ee-d6ba-2b27b29c597c/disk-0.vmdk. {{(pid=62814) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1168.246021] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7d7eb0e5-17d9-4462-ae04-c19cb5b929b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.246021] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294613, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072454} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.246021] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294612, 'name': CreateVM_Task, 'duration_secs': 0.36539} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.246701] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294614, 'name': ReconfigVM_Task, 'duration_secs': 0.419488} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.248682] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1168.248907] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1168.249179] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 881e9481-a5a9-489b-8abd-8efdaa239a11/881e9481-a5a9-489b-8abd-8efdaa239a11.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1168.250723] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaced170-f182-4469-b26d-b0232f825046 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.253311] env[62814]: DEBUG oslo_concurrency.lockutils [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.253491] env[62814]: DEBUG oslo_concurrency.lockutils [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1168.253806] env[62814]: DEBUG oslo_concurrency.lockutils [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1168.254032] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1cea85d5-bbe5-43f1-a385-fe08b105eac8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.256344] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6d7494c-549c-49a4-98f7-bd310cef22fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.275756] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 29468dc9-4254-4ac6-989c-c7fa0c4cc682/29468dc9-4254-4ac6-989c-c7fa0c4cc682.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1168.277916] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-258389ff-e867-448d-861b-efdbb71a18ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.292767] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1168.292767] env[62814]: value = "task-4294615" [ 1168.292767] env[62814]: _type = "Task" [ 1168.292767] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.293169] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1168.293169] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd2010-503b-40f6-5da2-3070a757e1fe" [ 1168.293169] env[62814]: _type = "Task" [ 1168.293169] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.310289] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1168.310289] env[62814]: value = "task-4294616" [ 1168.310289] env[62814]: _type = "Task" [ 1168.310289] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.310289] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd2010-503b-40f6-5da2-3070a757e1fe, 'name': SearchDatastore_Task, 'duration_secs': 0.009457} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.312909] env[62814]: DEBUG oslo_concurrency.lockutils [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1168.313235] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1168.316939] env[62814]: DEBUG oslo_concurrency.lockutils [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1168.316939] env[62814]: DEBUG oslo_concurrency.lockutils [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1168.316939] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1168.316939] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294615, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.321700] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6849c44c-d2b8-4a8d-833c-e113e8731662 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.326858] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294616, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.328512] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1168.328700] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1168.329491] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d45d032-1887-4c9b-8b7c-ce0b5db07cd1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.334652] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1168.334652] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522cbcee-7b81-7126-fe56-ada26550357f" [ 1168.334652] env[62814]: _type = "Task" [ 1168.334652] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.349076] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522cbcee-7b81-7126-fe56-ada26550357f, 'name': SearchDatastore_Task, 'duration_secs': 0.008369} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.349076] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24523f29-33b0-4f06-a680-fb61a5462cdd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.354908] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1168.354908] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a00dab-9f48-97d6-1a3d-ad538b6a4699" [ 1168.354908] env[62814]: _type = "Task" [ 1168.354908] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.362268] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a00dab-9f48-97d6-1a3d-ad538b6a4699, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.540147] env[62814]: INFO nova.compute.manager [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Shelve offloading [ 1168.604258] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1168.620230] env[62814]: DEBUG nova.network.neutron [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1168.708755] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1168.781438] env[62814]: DEBUG nova.network.neutron [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance_info_cache with network_info: [{"id": "580abcb7-b7f2-437a-8a84-738e83af0751", "address": "fa:16:3e:f1:6c:e8", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap580abcb7-b7", "ovs_interfaceid": "580abcb7-b7f2-437a-8a84-738e83af0751", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.797014] env[62814]: DEBUG nova.objects.instance [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lazy-loading 'flavor' on Instance uuid f00d414e-1dfd-43cf-9245-7cbeea8850b3 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1168.811285] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294615, 'name': Rename_Task, 'duration_secs': 0.145965} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.813243] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1168.813833] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6025e43c-8e88-4acb-bb0e-fd39bd50bc18 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.826401] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294616, 'name': ReconfigVM_Task, 'duration_secs': 0.325571} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.829135] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 29468dc9-4254-4ac6-989c-c7fa0c4cc682/29468dc9-4254-4ac6-989c-c7fa0c4cc682.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1168.831160] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1168.831160] env[62814]: value = "task-4294617" [ 1168.831160] env[62814]: _type = "Task" [ 1168.831160] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.831160] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f1a9a86-433d-48a3-a567-3fe2d13eb4cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.843939] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294617, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.846914] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1168.846914] env[62814]: value = "task-4294618" [ 1168.846914] env[62814]: _type = "Task" [ 1168.846914] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.855923] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294618, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1168.866812] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a00dab-9f48-97d6-1a3d-ad538b6a4699, 'name': SearchDatastore_Task, 'duration_secs': 0.008244} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1168.867109] env[62814]: DEBUG oslo_concurrency.lockutils [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1168.867469] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 53e13866-0174-4867-9fd9-2ea3f5ee399d/53e13866-0174-4867-9fd9-2ea3f5ee399d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1168.867748] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5ca14aa-dd66-4595-8f2e-d485a3f82281 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.876787] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1168.876787] env[62814]: value = "task-4294619" [ 1168.876787] env[62814]: _type = "Task" [ 1168.876787] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1168.886332] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294619, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.048785] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1169.049908] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1169.049908] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43dcbe77-bdd7-4c6c-a9a5-b63a16e3e23f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.060453] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1169.060453] env[62814]: value = "task-4294620" [ 1169.060453] env[62814]: _type = "Task" [ 1169.060453] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.073130] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1169.073411] env[62814]: DEBUG nova.compute.manager [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1169.074291] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a787c8-016c-4b90-9559-33daf9b5e891 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.082920] env[62814]: DEBUG oslo_concurrency.lockutils [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.083134] env[62814]: DEBUG oslo_concurrency.lockutils [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1169.083338] env[62814]: DEBUG nova.network.neutron [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1169.111810] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1169.111810] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.825s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1169.112668] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.404s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1169.114616] env[62814]: INFO nova.compute.claims [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1169.116772] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.116942] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Cleaning up deleted instances {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11865}} [ 1169.287178] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1169.287574] env[62814]: DEBUG nova.compute.manager [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Instance network_info: |[{"id": "580abcb7-b7f2-437a-8a84-738e83af0751", "address": "fa:16:3e:f1:6c:e8", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap580abcb7-b7", "ovs_interfaceid": "580abcb7-b7f2-437a-8a84-738e83af0751", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1169.288581] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:6c:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '605f83bd-808c-4b54-922e-54b14690987a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '580abcb7-b7f2-437a-8a84-738e83af0751', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1169.297067] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1169.299861] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1169.300445] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f982ceef-8f75-4aa8-8665-eb21a1f2cb54 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.325450] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c2b8bfd4-a598-466b-a462-c02353884c59 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.822s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1169.326763] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.278s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1169.331597] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1169.331597] env[62814]: value = "task-4294621" [ 1169.331597] env[62814]: _type = "Task" [ 1169.331597] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.349789] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294621, 'name': CreateVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.358782] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294617, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.367523] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294618, 'name': Rename_Task, 'duration_secs': 0.156574} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.367812] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1169.368508] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c4b247fb-0343-4d44-9145-88997b729133 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.375485] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1169.375485] env[62814]: value = "task-4294622" [ 1169.375485] env[62814]: _type = "Task" [ 1169.375485] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.386912] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294622, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.391583] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294619, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.456783] env[62814]: DEBUG nova.compute.manager [req-05b5f38a-a3a1-4317-8835-ab84c4e78b93 req-b5b83b86-3f3c-4792-8d79-afbe7c446dac service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Received event network-changed-580abcb7-b7f2-437a-8a84-738e83af0751 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1169.456926] env[62814]: DEBUG nova.compute.manager [req-05b5f38a-a3a1-4317-8835-ab84c4e78b93 req-b5b83b86-3f3c-4792-8d79-afbe7c446dac service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Refreshing instance network info cache due to event network-changed-580abcb7-b7f2-437a-8a84-738e83af0751. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1169.457185] env[62814]: DEBUG oslo_concurrency.lockutils [req-05b5f38a-a3a1-4317-8835-ab84c4e78b93 req-b5b83b86-3f3c-4792-8d79-afbe7c446dac service nova] Acquiring lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.457318] env[62814]: DEBUG oslo_concurrency.lockutils [req-05b5f38a-a3a1-4317-8835-ab84c4e78b93 req-b5b83b86-3f3c-4792-8d79-afbe7c446dac service nova] Acquired lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1169.457611] env[62814]: DEBUG nova.network.neutron [req-05b5f38a-a3a1-4317-8835-ab84c4e78b93 req-b5b83b86-3f3c-4792-8d79-afbe7c446dac service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Refreshing network info cache for port 580abcb7-b7f2-437a-8a84-738e83af0751 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1169.542206] env[62814]: DEBUG oslo_vmware.rw_handles [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Completed reading data from the image iterator. {{(pid=62814) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1169.542544] env[62814]: DEBUG oslo_vmware.rw_handles [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c367dc-ed09-93ee-d6ba-2b27b29c597c/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1169.543522] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83384f39-3e45-417f-bb44-131ed3f5e356 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.550673] env[62814]: DEBUG oslo_vmware.rw_handles [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c367dc-ed09-93ee-d6ba-2b27b29c597c/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1169.550864] env[62814]: DEBUG oslo_vmware.rw_handles [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c367dc-ed09-93ee-d6ba-2b27b29c597c/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1169.551145] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-568df82e-c5ba-4b24-91db-2a3653a2e179 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.636847] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] There are 58 instances to clean {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11874}} [ 1169.637051] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: eeb0255e-8c94-4564-af87-818cc8dee58e] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1169.789076] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d73d925e-39c6-4ea4-8ac3-a9a7df51124a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1169.789454] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d73d925e-39c6-4ea4-8ac3-a9a7df51124a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1169.789882] env[62814]: DEBUG nova.objects.instance [None req-d73d925e-39c6-4ea4-8ac3-a9a7df51124a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'flavor' on Instance uuid 2d7a18f3-456b-470c-a759-632e5d79a8f0 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1169.828352] env[62814]: DEBUG nova.network.neutron [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating instance_info_cache with network_info: [{"id": "a7104ea3-596d-4d10-952c-ac861087e580", "address": "fa:16:3e:4a:60:cd", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7104ea3-59", "ovs_interfaceid": "a7104ea3-596d-4d10-952c-ac861087e580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.832642] env[62814]: INFO nova.compute.manager [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Detaching volume 13a381a6-dd99-479c-a496-196303cb75b3 [ 1169.854441] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294621, 'name': CreateVM_Task, 'duration_secs': 0.367908} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.857329] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1169.858168] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294617, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.858294] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1169.858458] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1169.858770] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1169.859045] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87f91edf-5822-4942-80f8-c267f5559307 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.863843] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1169.863843] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524a32c6-5299-95e0-c48e-261969b3b8ac" [ 1169.863843] env[62814]: _type = "Task" [ 1169.863843] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.871743] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524a32c6-5299-95e0-c48e-261969b3b8ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.873160] env[62814]: INFO nova.virt.block_device [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Attempting to driver detach volume 13a381a6-dd99-479c-a496-196303cb75b3 from mountpoint /dev/sdb [ 1169.873473] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1169.873601] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845846', 'volume_id': '13a381a6-dd99-479c-a496-196303cb75b3', 'name': 'volume-13a381a6-dd99-479c-a496-196303cb75b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f00d414e-1dfd-43cf-9245-7cbeea8850b3', 'attached_at': '', 'detached_at': '', 'volume_id': '13a381a6-dd99-479c-a496-196303cb75b3', 'serial': '13a381a6-dd99-479c-a496-196303cb75b3'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1169.874379] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa8eaa0-1e6f-43c4-9fd3-3d95b80b67f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.903844] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876822cb-6163-474b-a3e8-42aee8eb787b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.906857] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294622, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.910520] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294619, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.528415} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.911200] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 53e13866-0174-4867-9fd9-2ea3f5ee399d/53e13866-0174-4867-9fd9-2ea3f5ee399d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1169.911565] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1169.911761] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8f5d7722-a780-49bb-b6f3-f1b3db174ae6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.916661] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5a69e2-a18b-4e6c-a11c-ae3b0292417b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.920082] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1169.920082] env[62814]: value = "task-4294623" [ 1169.920082] env[62814]: _type = "Task" [ 1169.920082] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.942772] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd05adbd-962e-4c63-84f8-43f5e930596c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.949886] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294623, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.965121] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] The volume has not been displaced from its original location: [datastore1] volume-13a381a6-dd99-479c-a496-196303cb75b3/volume-13a381a6-dd99-479c-a496-196303cb75b3.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1169.972096] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Reconfiguring VM instance instance-00000063 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1169.972096] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3309578d-bf86-4332-8ce5-7c7a1a6648e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.991409] env[62814]: DEBUG oslo_vmware.api [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1169.991409] env[62814]: value = "task-4294624" [ 1169.991409] env[62814]: _type = "Task" [ 1169.991409] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.999303] env[62814]: DEBUG oslo_vmware.api [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294624, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.012151] env[62814]: DEBUG oslo_vmware.rw_handles [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c367dc-ed09-93ee-d6ba-2b27b29c597c/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1170.012399] env[62814]: INFO nova.virt.vmwareapi.images [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Downloaded image file data dc3cbb42-9340-472a-b42a-4fb9bea7421b [ 1170.013256] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-785c6019-3d04-4d4d-a9a0-2a510c25fe7f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.031419] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25257b17-a797-4b30-9c51-75cbdd42e910 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.132995] env[62814]: INFO nova.virt.vmwareapi.images [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] The imported VM was unregistered [ 1170.135635] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Caching image {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1170.135883] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Creating directory with path [datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1170.136173] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44b27e4b-dc1d-4f4e-9b38-43780ffd1609 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.143172] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: e75c8039-9ff8-45f2-8770-894f2307cc14] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1170.149291] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Created directory with path [datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1170.149291] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2/OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2.vmdk to [datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b/dc3cbb42-9340-472a-b42a-4fb9bea7421b.vmdk. {{(pid=62814) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1170.149920] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-93c5765e-fcbd-4c97-8dd0-861ef20a534c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.164554] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1170.164554] env[62814]: value = "task-4294626" [ 1170.164554] env[62814]: _type = "Task" [ 1170.164554] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.173682] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294626, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.225666] env[62814]: DEBUG nova.network.neutron [req-05b5f38a-a3a1-4317-8835-ab84c4e78b93 req-b5b83b86-3f3c-4792-8d79-afbe7c446dac service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updated VIF entry in instance network info cache for port 580abcb7-b7f2-437a-8a84-738e83af0751. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1170.225666] env[62814]: DEBUG nova.network.neutron [req-05b5f38a-a3a1-4317-8835-ab84c4e78b93 req-b5b83b86-3f3c-4792-8d79-afbe7c446dac service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance_info_cache with network_info: [{"id": "580abcb7-b7f2-437a-8a84-738e83af0751", "address": "fa:16:3e:f1:6c:e8", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap580abcb7-b7", "ovs_interfaceid": "580abcb7-b7f2-437a-8a84-738e83af0751", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1170.293516] env[62814]: DEBUG nova.objects.instance [None req-d73d925e-39c6-4ea4-8ac3-a9a7df51124a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'pci_requests' on Instance uuid 2d7a18f3-456b-470c-a759-632e5d79a8f0 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1170.330653] env[62814]: DEBUG oslo_concurrency.lockutils [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1170.340376] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b19fc64-17a1-4380-944d-c6dafed4e9ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.350048] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3745f83f-31c9-42a5-ac14-ce9a23547abe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.356544] env[62814]: DEBUG oslo_vmware.api [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294617, 'name': PowerOnVM_Task, 'duration_secs': 1.102677} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.357478] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1170.357560] env[62814]: INFO nova.compute.manager [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Took 8.63 seconds to spawn the instance on the hypervisor. [ 1170.357766] env[62814]: DEBUG nova.compute.manager [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1170.358594] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66a33e3-7585-4691-9697-8029c4fe9931 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.398314] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9462de44-b0a2-4098-a93f-f078aeb052b8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.411903] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]524a32c6-5299-95e0-c48e-261969b3b8ac, 'name': SearchDatastore_Task, 'duration_secs': 0.022077} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.417551] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1170.417845] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1170.418163] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.418320] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1170.418627] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1170.418887] env[62814]: DEBUG oslo_vmware.api [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294622, 'name': PowerOnVM_Task, 'duration_secs': 0.848978} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.419215] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05bcc0b9-e6e1-4b28-bd68-4782566fed71 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.421789] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1170.422049] env[62814]: DEBUG nova.compute.manager [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1170.425654] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0dcdd39-9b76-4bd4-a976-9d77785bd2eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.428266] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a3828b-e894-4b9e-b0ee-e64d109eddbb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.454449] env[62814]: DEBUG nova.compute.provider_tree [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.456073] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294623, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072722} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.458146] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1170.458588] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1170.458777] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1170.462659] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-732149fe-aedd-48f5-b818-a03cd4d8e2bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.465553] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2124ad3e-01c4-419c-a9cd-503ffc625eaa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.490474] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 53e13866-0174-4867-9fd9-2ea3f5ee399d/53e13866-0174-4867-9fd9-2ea3f5ee399d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1170.493065] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4e17174f-8889-43c5-a2b3-449fc4646dc0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.509936] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1170.509936] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523628df-61de-5984-07bf-6f442726e560" [ 1170.509936] env[62814]: _type = "Task" [ 1170.509936] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.521149] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1170.521149] env[62814]: value = "task-4294627" [ 1170.521149] env[62814]: _type = "Task" [ 1170.521149] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.525363] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523628df-61de-5984-07bf-6f442726e560, 'name': SearchDatastore_Task, 'duration_secs': 0.048908} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.531992] env[62814]: DEBUG oslo_vmware.api [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294624, 'name': ReconfigVM_Task, 'duration_secs': 0.222981} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.532618] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-071d6cc5-c5ae-40a1-82ce-858b91f8e18a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.535302] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Reconfigured VM instance instance-00000063 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1170.540640] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac08350c-e410-4a36-b44a-3e952c47a918 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.554931] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294627, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.557631] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1170.557631] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522fbee4-3220-e813-3fca-a5afc143592f" [ 1170.557631] env[62814]: _type = "Task" [ 1170.557631] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.576818] env[62814]: DEBUG oslo_vmware.api [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1170.576818] env[62814]: value = "task-4294628" [ 1170.576818] env[62814]: _type = "Task" [ 1170.576818] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.594663] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522fbee4-3220-e813-3fca-a5afc143592f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.594663] env[62814]: DEBUG oslo_vmware.api [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294628, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.648411] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 08c20588-2988-4976-8631-53cf447b4877] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1170.676804] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294626, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.727566] env[62814]: DEBUG oslo_concurrency.lockutils [req-05b5f38a-a3a1-4317-8835-ab84c4e78b93 req-b5b83b86-3f3c-4792-8d79-afbe7c446dac service nova] Releasing lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1170.796496] env[62814]: DEBUG nova.objects.base [None req-d73d925e-39c6-4ea4-8ac3-a9a7df51124a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Object Instance<2d7a18f3-456b-470c-a759-632e5d79a8f0> lazy-loaded attributes: flavor,pci_requests {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1170.796747] env[62814]: DEBUG nova.network.neutron [None req-d73d925e-39c6-4ea4-8ac3-a9a7df51124a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1170.878055] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1170.878980] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99af603-3d47-49c3-9423-28192a3b7b95 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.891815] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1170.892116] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6137978d-e209-4896-b811-a1bce04ad6e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.904774] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d73d925e-39c6-4ea4-8ac3-a9a7df51124a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.115s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1170.914239] env[62814]: INFO nova.compute.manager [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Took 13.56 seconds to build instance. [ 1170.958334] env[62814]: DEBUG nova.scheduler.client.report [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1170.968424] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1170.998327] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1170.998564] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1170.998749] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleting the datastore file [datastore2] 1f13051e-e51b-4981-9445-d5420c3c1818 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1170.999047] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a0fe2235-40c1-4422-915a-302e9b86c902 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.008749] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1171.008749] env[62814]: value = "task-4294630" [ 1171.008749] env[62814]: _type = "Task" [ 1171.008749] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.021210] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.036785] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294627, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.072966] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522fbee4-3220-e813-3fca-a5afc143592f, 'name': SearchDatastore_Task, 'duration_secs': 0.082503} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.073789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1171.074155] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 40e3c00b-2129-476e-bc67-a0ef13781990/40e3c00b-2129-476e-bc67-a0ef13781990.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1171.074471] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2046360c-192b-48c2-85b4-420087e81946 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.079882] env[62814]: DEBUG oslo_vmware.api [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294628, 'name': ReconfigVM_Task, 'duration_secs': 0.428771} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.080578] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845846', 'volume_id': '13a381a6-dd99-479c-a496-196303cb75b3', 'name': 'volume-13a381a6-dd99-479c-a496-196303cb75b3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'f00d414e-1dfd-43cf-9245-7cbeea8850b3', 'attached_at': '', 'detached_at': '', 'volume_id': '13a381a6-dd99-479c-a496-196303cb75b3', 'serial': '13a381a6-dd99-479c-a496-196303cb75b3'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1171.087810] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1171.087810] env[62814]: value = "task-4294631" [ 1171.087810] env[62814]: _type = "Task" [ 1171.087810] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.098158] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294631, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.155075] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 201cf7be-f81f-4eba-ae65-a370ad8e48c3] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1171.177092] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294626, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.420445] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eae3daa5-709c-4640-a117-2ddba2af74b4 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "881e9481-a5a9-489b-8abd-8efdaa239a11" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.072s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.467382] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.355s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.468048] env[62814]: DEBUG nova.compute.manager [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1171.470829] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.502s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.471095] env[62814]: DEBUG nova.objects.instance [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1171.522945] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.537252] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294627, 'name': ReconfigVM_Task, 'duration_secs': 0.878769} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1171.537630] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 53e13866-0174-4867-9fd9-2ea3f5ee399d/53e13866-0174-4867-9fd9-2ea3f5ee399d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1171.538971] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a11b338b-17c5-4460-b44d-15edc9fbb0cc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.548310] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1171.548310] env[62814]: value = "task-4294632" [ 1171.548310] env[62814]: _type = "Task" [ 1171.548310] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.559689] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294632, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.600586] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294631, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.638967] env[62814]: DEBUG nova.objects.instance [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lazy-loading 'flavor' on Instance uuid f00d414e-1dfd-43cf-9245-7cbeea8850b3 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1171.658746] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 4429ff0c-5353-4346-8bef-2c45d251f1ea] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1171.677880] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294626, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.769134] env[62814]: DEBUG nova.compute.manager [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received event network-vif-unplugged-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1171.769403] env[62814]: DEBUG oslo_concurrency.lockutils [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] Acquiring lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1171.769654] env[62814]: DEBUG oslo_concurrency.lockutils [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] Lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1171.769839] env[62814]: DEBUG oslo_concurrency.lockutils [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] Lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1171.770052] env[62814]: DEBUG nova.compute.manager [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] No waiting events found dispatching network-vif-unplugged-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1171.770286] env[62814]: WARNING nova.compute.manager [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received unexpected event network-vif-unplugged-a7104ea3-596d-4d10-952c-ac861087e580 for instance with vm_state shelved and task_state shelving_offloading. [ 1171.770460] env[62814]: DEBUG nova.compute.manager [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received event network-changed-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1171.770616] env[62814]: DEBUG nova.compute.manager [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Refreshing instance network info cache due to event network-changed-a7104ea3-596d-4d10-952c-ac861087e580. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1171.770804] env[62814]: DEBUG oslo_concurrency.lockutils [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] Acquiring lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.770940] env[62814]: DEBUG oslo_concurrency.lockutils [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] Acquired lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1171.771120] env[62814]: DEBUG nova.network.neutron [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Refreshing network info cache for port a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1171.977182] env[62814]: DEBUG nova.compute.utils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1171.982341] env[62814]: DEBUG nova.compute.manager [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1171.982522] env[62814]: DEBUG nova.network.neutron [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1172.031140] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.043583] env[62814]: DEBUG nova.policy [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd71116c0cfd4b509e4070ecc84613f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd82993ef7dfa4d2f8f39db3577dd321a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1172.060315] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294632, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.102954] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294631, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.162375] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 7e849b23-226d-4a1d-a85c-cf96964ce034] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1172.187293] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294626, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.435357] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "29468dc9-4254-4ac6-989c-c7fa0c4cc682" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.435660] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "29468dc9-4254-4ac6-989c-c7fa0c4cc682" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.435856] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "29468dc9-4254-4ac6-989c-c7fa0c4cc682-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.436048] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "29468dc9-4254-4ac6-989c-c7fa0c4cc682-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.436228] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "29468dc9-4254-4ac6-989c-c7fa0c4cc682-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.439538] env[62814]: INFO nova.compute.manager [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Terminating instance [ 1172.482987] env[62814]: DEBUG nova.compute.manager [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1172.489811] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a0ab2966-17c2-4143-a18f-18837833491d tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.525068] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294630, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.532635] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "881e9481-a5a9-489b-8abd-8efdaa239a11" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.532903] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "881e9481-a5a9-489b-8abd-8efdaa239a11" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.533059] env[62814]: DEBUG nova.compute.manager [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1172.534407] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6e61458-e2cf-4efd-be02-ef47b45d5ccf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.550574] env[62814]: DEBUG nova.compute.manager [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1172.550574] env[62814]: DEBUG nova.objects.instance [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lazy-loading 'flavor' on Instance uuid 881e9481-a5a9-489b-8abd-8efdaa239a11 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.560909] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294632, 'name': Rename_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.603063] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294631, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.655171] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc4ed5f2-2e76-4991-9769-c294ae1e57dc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.328s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1172.666504] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 44ea319c-6ea0-456a-bee6-42133a25d8c1] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1172.679543] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294626, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.828906] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1172.829240] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1172.829575] env[62814]: DEBUG nova.objects.instance [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'flavor' on Instance uuid 2d7a18f3-456b-470c-a759-632e5d79a8f0 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.944044] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "refresh_cache-29468dc9-4254-4ac6-989c-c7fa0c4cc682" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.944217] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired lock "refresh_cache-29468dc9-4254-4ac6-989c-c7fa0c4cc682" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1172.944392] env[62814]: DEBUG nova.network.neutron [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1173.010447] env[62814]: DEBUG nova.network.neutron [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Successfully created port: 461b5db0-7363-4a49-8a69-75f620fd7861 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1173.023398] env[62814]: DEBUG oslo_vmware.api [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294630, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.582917} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.024044] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.024044] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1173.024238] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1173.028327] env[62814]: DEBUG nova.network.neutron [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updated VIF entry in instance network info cache for port a7104ea3-596d-4d10-952c-ac861087e580. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1173.028657] env[62814]: DEBUG nova.network.neutron [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating instance_info_cache with network_info: [{"id": "a7104ea3-596d-4d10-952c-ac861087e580", "address": "fa:16:3e:4a:60:cd", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapa7104ea3-59", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.051443] env[62814]: INFO nova.scheduler.client.report [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleted allocations for instance 1f13051e-e51b-4981-9445-d5420c3c1818 [ 1173.065813] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294632, 'name': Rename_Task, 'duration_secs': 1.325391} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.066081] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1173.066328] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7fd1b69-c598-42be-87ec-93b16e48b5ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.073361] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1173.073361] env[62814]: value = "task-4294633" [ 1173.073361] env[62814]: _type = "Task" [ 1173.073361] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.081996] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294633, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.101843] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294631, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.812739} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.102119] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 40e3c00b-2129-476e-bc67-a0ef13781990/40e3c00b-2129-476e-bc67-a0ef13781990.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1173.102339] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1173.102639] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3f7f5e9b-1c12-421a-8d49-41c0da9393a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.109177] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1173.109177] env[62814]: value = "task-4294634" [ 1173.109177] env[62814]: _type = "Task" [ 1173.109177] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.121495] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294634, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.172673] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: d5df1d0e-55e6-452e-882a-053a83250b39] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1173.180151] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294626, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.547521} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.180433] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2/OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2.vmdk to [datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b/dc3cbb42-9340-472a-b42a-4fb9bea7421b.vmdk. [ 1173.180626] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Cleaning up location [datastore2] OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1173.180789] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_61b6b05e-1d30-4588-a19e-38beedef52a2 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1173.181065] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-81f13a57-179a-42f1-a289-9a220586f524 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.187590] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1173.187590] env[62814]: value = "task-4294635" [ 1173.187590] env[62814]: _type = "Task" [ 1173.187590] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.195664] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294635, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.434035] env[62814]: DEBUG nova.objects.instance [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'pci_requests' on Instance uuid 2d7a18f3-456b-470c-a759-632e5d79a8f0 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.469008] env[62814]: DEBUG nova.network.neutron [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1173.498725] env[62814]: DEBUG nova.compute.manager [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1173.528162] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1173.528468] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1173.528655] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1173.528884] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1173.529173] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1173.529404] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1173.529717] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1173.529919] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1173.530117] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1173.530304] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1173.530523] env[62814]: DEBUG nova.virt.hardware [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1173.531479] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9897ac8-a2fd-43b5-8546-bfd8f584c565 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.534943] env[62814]: DEBUG oslo_concurrency.lockutils [req-e21815f6-9f04-429c-8486-3e3191c0e090 req-2d7dcbfd-2c38-433c-88f5-9b1d293ec45e service nova] Releasing lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1173.540419] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54058f7b-ebb3-412f-8996-e7a466ea0058 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.546677] env[62814]: DEBUG nova.network.neutron [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1173.560399] env[62814]: DEBUG oslo_concurrency.lockutils [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1173.560713] env[62814]: DEBUG oslo_concurrency.lockutils [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1173.560966] env[62814]: DEBUG nova.objects.instance [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'resources' on Instance uuid 1f13051e-e51b-4981-9445-d5420c3c1818 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1173.563182] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Releasing lock "refresh_cache-29468dc9-4254-4ac6-989c-c7fa0c4cc682" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1173.563674] env[62814]: DEBUG nova.compute.manager [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1173.563907] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1173.564398] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1173.565455] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3330de5e-6530-48d7-8752-4301d80b9178 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.568543] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-da95f3fb-3104-4bd7-89fa-33c5fa869447 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.575744] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1173.579672] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db55ef00-27ce-4901-9368-bc91f1309a7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.581307] env[62814]: DEBUG oslo_vmware.api [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1173.581307] env[62814]: value = "task-4294636" [ 1173.581307] env[62814]: _type = "Task" [ 1173.581307] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.588571] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294633, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.590237] env[62814]: DEBUG oslo_vmware.api [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1173.590237] env[62814]: value = "task-4294637" [ 1173.590237] env[62814]: _type = "Task" [ 1173.590237] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.596766] env[62814]: DEBUG oslo_vmware.api [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294636, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.601531] env[62814]: DEBUG oslo_vmware.api [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294637, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.618197] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294634, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072447} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.618502] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1173.619310] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62984560-8b3f-4bbb-9855-f25010e7e9e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.642144] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 40e3c00b-2129-476e-bc67-a0ef13781990/40e3c00b-2129-476e-bc67-a0ef13781990.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1173.642996] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85e70e93-a505-4adc-8237-b308f75ac96b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.666820] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1173.666820] env[62814]: value = "task-4294638" [ 1173.666820] env[62814]: _type = "Task" [ 1173.666820] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.676423] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 19b53797-6fc9-4bb0-ab3f-9d2941b11cb3] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1173.682989] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294638, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.698190] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294635, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033539} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.698469] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1173.698636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "[datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b/dc3cbb42-9340-472a-b42a-4fb9bea7421b.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1173.698992] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b/dc3cbb42-9340-472a-b42a-4fb9bea7421b.vmdk to [datastore2] edcdb1e5-ed9d-49e6-97e6-ea7629682547/edcdb1e5-ed9d-49e6-97e6-ea7629682547.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1173.699438] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-650732e2-0fe2-46e4-8594-11744145c0d6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.707672] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1173.707672] env[62814]: value = "task-4294639" [ 1173.707672] env[62814]: _type = "Task" [ 1173.707672] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.716189] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294639, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.939245] env[62814]: DEBUG nova.objects.base [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Object Instance<2d7a18f3-456b-470c-a759-632e5d79a8f0> lazy-loaded attributes: flavor,pci_requests {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1173.939492] env[62814]: DEBUG nova.network.neutron [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1173.986703] env[62814]: DEBUG nova.policy [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5779eaa60ec44b2b80660e521b598eb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10018060297c409d82935ad626cc54c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1174.069538] env[62814]: DEBUG nova.objects.instance [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'numa_topology' on Instance uuid 1f13051e-e51b-4981-9445-d5420c3c1818 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1174.088916] env[62814]: DEBUG oslo_vmware.api [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294633, 'name': PowerOnVM_Task, 'duration_secs': 0.862615} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.089751] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1174.089875] env[62814]: DEBUG nova.compute.manager [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1174.090667] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4b3b77-9928-4243-9aa2-e46a7a3aec50 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.097343] env[62814]: DEBUG oslo_vmware.api [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294636, 'name': PowerOffVM_Task, 'duration_secs': 0.249855} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.100957] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1174.101197] env[62814]: DEBUG nova.compute.manager [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1174.101986] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ebe353-90a4-4809-a07b-dd05a5d18e31 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.117669] env[62814]: DEBUG oslo_vmware.api [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294637, 'name': PowerOffVM_Task, 'duration_secs': 0.166495} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.119603] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1174.119802] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1174.122091] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-14e84e1d-73e4-4be0-89b5-2a05bd2f6dfc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.149718] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1174.150227] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1174.150227] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Deleting the datastore file [datastore2] 29468dc9-4254-4ac6-989c-c7fa0c4cc682 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1174.150227] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53061ba8-24f5-45a1-a3d8-32a06c02a8b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.156905] env[62814]: DEBUG oslo_vmware.api [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1174.156905] env[62814]: value = "task-4294641" [ 1174.156905] env[62814]: _type = "Task" [ 1174.156905] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.162074] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1174.162392] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.162567] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1174.162748] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1174.162951] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1174.168431] env[62814]: DEBUG oslo_vmware.api [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294641, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.168931] env[62814]: INFO nova.compute.manager [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Terminating instance [ 1174.180200] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294638, 'name': ReconfigVM_Task, 'duration_secs': 0.345137} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.180519] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 40e3c00b-2129-476e-bc67-a0ef13781990/40e3c00b-2129-476e-bc67-a0ef13781990.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1174.181178] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2fb83642-e984-4b16-aef6-f7616b0c8f48 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.183635] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: cd037f6e-fddd-4389-b6d0-144b798537bf] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1174.190888] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1174.190888] env[62814]: value = "task-4294642" [ 1174.190888] env[62814]: _type = "Task" [ 1174.190888] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.200409] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294642, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.218893] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294639, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.332610] env[62814]: DEBUG nova.network.neutron [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Successfully created port: 88931aee-29b6-4f58-9ee6-0210612c8211 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1174.572486] env[62814]: DEBUG nova.objects.base [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Object Instance<1f13051e-e51b-4981-9445-d5420c3c1818> lazy-loaded attributes: resources,numa_topology {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1174.617375] env[62814]: DEBUG oslo_concurrency.lockutils [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1174.627839] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4e36d19a-081e-4ab3-a6f9-cc31955a2ad1 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "881e9481-a5a9-489b-8abd-8efdaa239a11" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.095s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1174.672342] env[62814]: DEBUG oslo_vmware.api [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294641, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.677883] env[62814]: DEBUG nova.compute.manager [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1174.678219] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1174.679528] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62783145-248e-4320-970e-b9c192ae68a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.689269] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 0c6c598f-ef5a-4e91-b811-cd3d8f072647] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1174.693953] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1174.697412] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e471408-8b5d-4a9c-9648-6d23ad404c32 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.709586] env[62814]: DEBUG oslo_vmware.api [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1174.709586] env[62814]: value = "task-4294643" [ 1174.709586] env[62814]: _type = "Task" [ 1174.709586] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.718236] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294642, 'name': Rename_Task, 'duration_secs': 0.495486} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.725443] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1174.727054] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3361c910-5d9b-424f-9a73-f5633733f1ef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.733799] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294639, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.737109] env[62814]: DEBUG oslo_vmware.api [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294643, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.741373] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1174.741373] env[62814]: value = "task-4294644" [ 1174.741373] env[62814]: _type = "Task" [ 1174.741373] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.755199] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294644, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.831549] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e2893b-50f0-4082-9b23-01d8e5dc2ed7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.844258] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bfbfac-017e-4ce9-bcd1-8266ef8c5563 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.882028] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1a939b0-2a7a-4f21-955d-4dbfb475010b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.893466] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9915a352-81b8-456b-ac1f-5939d471e015 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.914468] env[62814]: DEBUG nova.compute.provider_tree [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1174.987489] env[62814]: DEBUG nova.network.neutron [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Successfully updated port: 461b5db0-7363-4a49-8a69-75f620fd7861 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1175.074405] env[62814]: DEBUG nova.compute.manager [req-0ea711b8-3913-4299-8f72-8ddad62afd88 req-37948af8-348b-4812-a5d3-e60fb0e908d6 service nova] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Received event network-vif-plugged-461b5db0-7363-4a49-8a69-75f620fd7861 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1175.074639] env[62814]: DEBUG oslo_concurrency.lockutils [req-0ea711b8-3913-4299-8f72-8ddad62afd88 req-37948af8-348b-4812-a5d3-e60fb0e908d6 service nova] Acquiring lock "bf2ba498-7c47-4985-a84a-475037deec55-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1175.074847] env[62814]: DEBUG oslo_concurrency.lockutils [req-0ea711b8-3913-4299-8f72-8ddad62afd88 req-37948af8-348b-4812-a5d3-e60fb0e908d6 service nova] Lock "bf2ba498-7c47-4985-a84a-475037deec55-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1175.075018] env[62814]: DEBUG oslo_concurrency.lockutils [req-0ea711b8-3913-4299-8f72-8ddad62afd88 req-37948af8-348b-4812-a5d3-e60fb0e908d6 service nova] Lock "bf2ba498-7c47-4985-a84a-475037deec55-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1175.075189] env[62814]: DEBUG nova.compute.manager [req-0ea711b8-3913-4299-8f72-8ddad62afd88 req-37948af8-348b-4812-a5d3-e60fb0e908d6 service nova] [instance: bf2ba498-7c47-4985-a84a-475037deec55] No waiting events found dispatching network-vif-plugged-461b5db0-7363-4a49-8a69-75f620fd7861 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1175.075379] env[62814]: WARNING nova.compute.manager [req-0ea711b8-3913-4299-8f72-8ddad62afd88 req-37948af8-348b-4812-a5d3-e60fb0e908d6 service nova] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Received unexpected event network-vif-plugged-461b5db0-7363-4a49-8a69-75f620fd7861 for instance with vm_state building and task_state spawning. [ 1175.170166] env[62814]: DEBUG oslo_vmware.api [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294641, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.196994] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: b4ee4cee-3298-4955-8375-8ca8c04b2f9f] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1175.228403] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294639, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.233913] env[62814]: DEBUG oslo_vmware.api [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294643, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.252809] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294644, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.416037] env[62814]: DEBUG nova.scheduler.client.report [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1175.490728] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "refresh_cache-bf2ba498-7c47-4985-a84a-475037deec55" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.490919] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "refresh_cache-bf2ba498-7c47-4985-a84a-475037deec55" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1175.491710] env[62814]: DEBUG nova.network.neutron [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1175.668962] env[62814]: DEBUG oslo_vmware.api [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294641, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.198784} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.669309] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1175.669525] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1175.669718] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1175.669896] env[62814]: INFO nova.compute.manager [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Took 2.11 seconds to destroy the instance on the hypervisor. [ 1175.670177] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1175.670424] env[62814]: DEBUG nova.compute.manager [-] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1175.670551] env[62814]: DEBUG nova.network.neutron [-] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1175.698199] env[62814]: DEBUG nova.network.neutron [-] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1175.700292] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 1c86646f-6f38-4f8d-bea0-8a6b93680aba] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1175.721064] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294639, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.730239] env[62814]: DEBUG oslo_vmware.api [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294643, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.754777] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294644, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.924417] env[62814]: DEBUG oslo_concurrency.lockutils [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.363s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1175.927538] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "881e9481-a5a9-489b-8abd-8efdaa239a11" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1175.927786] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "881e9481-a5a9-489b-8abd-8efdaa239a11" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1175.927998] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "881e9481-a5a9-489b-8abd-8efdaa239a11-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1175.928281] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "881e9481-a5a9-489b-8abd-8efdaa239a11-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1175.928537] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "881e9481-a5a9-489b-8abd-8efdaa239a11-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1175.930425] env[62814]: DEBUG oslo_concurrency.lockutils [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.313s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1175.930671] env[62814]: DEBUG nova.objects.instance [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1175.934486] env[62814]: INFO nova.compute.manager [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Terminating instance [ 1176.012288] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "1f13051e-e51b-4981-9445-d5420c3c1818" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1176.012288] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "53e13866-0174-4867-9fd9-2ea3f5ee399d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1176.012415] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "53e13866-0174-4867-9fd9-2ea3f5ee399d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.016022] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "53e13866-0174-4867-9fd9-2ea3f5ee399d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1176.016022] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "53e13866-0174-4867-9fd9-2ea3f5ee399d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.016022] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "53e13866-0174-4867-9fd9-2ea3f5ee399d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.016022] env[62814]: INFO nova.compute.manager [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Terminating instance [ 1176.033575] env[62814]: DEBUG nova.network.neutron [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1176.202652] env[62814]: DEBUG nova.network.neutron [-] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.204152] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: cfab341e-57a1-48b4-9b6b-ceecf28c223b] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1176.221119] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294639, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.503158} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.225083] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/dc3cbb42-9340-472a-b42a-4fb9bea7421b/dc3cbb42-9340-472a-b42a-4fb9bea7421b.vmdk to [datastore2] edcdb1e5-ed9d-49e6-97e6-ea7629682547/edcdb1e5-ed9d-49e6-97e6-ea7629682547.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1176.226157] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2557e2-79a5-488a-abb0-d6d086200bb8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.233493] env[62814]: DEBUG oslo_vmware.api [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294643, 'name': PowerOffVM_Task, 'duration_secs': 1.033271} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.243900] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1176.244191] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1176.257705] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] edcdb1e5-ed9d-49e6-97e6-ea7629682547/edcdb1e5-ed9d-49e6-97e6-ea7629682547.vmdk or device None with type streamOptimized {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1176.258818] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a58f6864-2b02-490c-b48e-83c996bb542a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.260742] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7bffbb6e-9b3f-4493-8c74-a13b5441bb5c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.287176] env[62814]: DEBUG oslo_vmware.api [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294644, 'name': PowerOnVM_Task, 'duration_secs': 1.131321} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.288691] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1176.288916] env[62814]: INFO nova.compute.manager [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Took 9.46 seconds to spawn the instance on the hypervisor. [ 1176.289113] env[62814]: DEBUG nova.compute.manager [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1176.289448] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1176.289448] env[62814]: value = "task-4294646" [ 1176.289448] env[62814]: _type = "Task" [ 1176.289448] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.290320] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d08dcab9-bd28-43a7-82c5-1d28069824b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.302121] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294646, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.303892] env[62814]: DEBUG nova.network.neutron [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Updating instance_info_cache with network_info: [{"id": "461b5db0-7363-4a49-8a69-75f620fd7861", "address": "fa:16:3e:23:01:29", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461b5db0-73", "ovs_interfaceid": "461b5db0-7363-4a49-8a69-75f620fd7861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.343051] env[62814]: DEBUG nova.network.neutron [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Successfully updated port: 88931aee-29b6-4f58-9ee6-0210612c8211 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1176.378735] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1176.379146] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1176.379605] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleting the datastore file [datastore2] f00d414e-1dfd-43cf-9245-7cbeea8850b3 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1176.379605] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5792c9b5-d0f3-4baf-b554-ee458441f6a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.387308] env[62814]: DEBUG oslo_vmware.api [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1176.387308] env[62814]: value = "task-4294647" [ 1176.387308] env[62814]: _type = "Task" [ 1176.387308] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.395536] env[62814]: DEBUG oslo_vmware.api [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294647, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.441339] env[62814]: DEBUG oslo_concurrency.lockutils [None req-63019387-c11a-4251-9d3a-f657a48d5d8a tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.339s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1176.443486] env[62814]: DEBUG nova.compute.manager [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1176.443486] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1176.443917] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.432s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1176.444230] env[62814]: INFO nova.compute.manager [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Unshelving [ 1176.447094] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69058341-7621-481a-84a8-330701df7ad0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.459619] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1176.460015] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8b951c79-6283-4b53-817f-262f45c0d1f2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.520088] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "refresh_cache-53e13866-0174-4867-9fd9-2ea3f5ee399d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.520439] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquired lock "refresh_cache-53e13866-0174-4867-9fd9-2ea3f5ee399d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1176.520846] env[62814]: DEBUG nova.network.neutron [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1176.528192] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1176.528530] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1176.528839] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleting the datastore file [datastore2] 881e9481-a5a9-489b-8abd-8efdaa239a11 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1176.529239] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cdb1eac8-9c73-4fe1-a34b-425625cda852 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.537409] env[62814]: DEBUG oslo_vmware.api [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1176.537409] env[62814]: value = "task-4294649" [ 1176.537409] env[62814]: _type = "Task" [ 1176.537409] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.549952] env[62814]: DEBUG oslo_vmware.api [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294649, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.708071] env[62814]: INFO nova.compute.manager [-] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Took 1.04 seconds to deallocate network for instance. [ 1176.710265] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 88ec5aba-f38d-4c75-af29-e3df3aa49640] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1176.802961] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294646, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.808488] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "refresh_cache-bf2ba498-7c47-4985-a84a-475037deec55" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1176.808788] env[62814]: DEBUG nova.compute.manager [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Instance network_info: |[{"id": "461b5db0-7363-4a49-8a69-75f620fd7861", "address": "fa:16:3e:23:01:29", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461b5db0-73", "ovs_interfaceid": "461b5db0-7363-4a49-8a69-75f620fd7861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1176.812660] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:01:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '461b5db0-7363-4a49-8a69-75f620fd7861', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1176.820062] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1176.822036] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1176.822480] env[62814]: INFO nova.compute.manager [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Took 18.26 seconds to build instance. [ 1176.823295] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06cc7d4c-9a99-4ebe-a703-db714612af8b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.843717] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1176.843717] env[62814]: value = "task-4294650" [ 1176.843717] env[62814]: _type = "Task" [ 1176.843717] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1176.847104] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.847368] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1176.847583] env[62814]: DEBUG nova.network.neutron [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1176.854201] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294650, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1176.897628] env[62814]: DEBUG oslo_vmware.api [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294647, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166564} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.897881] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1176.898168] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1176.898375] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1176.898554] env[62814]: INFO nova.compute.manager [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1176.898861] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1176.899082] env[62814]: DEBUG nova.compute.manager [-] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1176.899181] env[62814]: DEBUG nova.network.neutron [-] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1176.947155] env[62814]: DEBUG oslo_concurrency.lockutils [None req-165918fc-7e53-4723-9c09-26d334033cd5 tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.046727] env[62814]: DEBUG nova.network.neutron [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1177.052098] env[62814]: DEBUG oslo_vmware.api [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294649, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145984} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.052328] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1177.052569] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1177.052684] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1177.052865] env[62814]: INFO nova.compute.manager [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1177.053133] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1177.053334] env[62814]: DEBUG nova.compute.manager [-] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1177.053495] env[62814]: DEBUG nova.network.neutron [-] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1177.188427] env[62814]: DEBUG nova.network.neutron [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.216330] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: a07b964f-fe88-4c88-b6f4-61ed0973716c] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1177.219491] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1177.219823] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1177.219958] env[62814]: DEBUG nova.objects.instance [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lazy-loading 'resources' on Instance uuid 29468dc9-4254-4ac6-989c-c7fa0c4cc682 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1177.306153] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294646, 'name': ReconfigVM_Task, 'duration_secs': 0.901236} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.306153] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Reconfigured VM instance instance-00000050 to attach disk [datastore2] edcdb1e5-ed9d-49e6-97e6-ea7629682547/edcdb1e5-ed9d-49e6-97e6-ea7629682547.vmdk or device None with type streamOptimized {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1177.307094] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57046adc-e06d-489f-a56b-b90e7295467c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.314211] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1177.314211] env[62814]: value = "task-4294651" [ 1177.314211] env[62814]: _type = "Task" [ 1177.314211] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.324976] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294651, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.338720] env[62814]: DEBUG oslo_concurrency.lockutils [None req-96f4226c-c031-4554-9457-2c9fe344fb3b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "40e3c00b-2129-476e-bc67-a0ef13781990" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.792s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1177.359192] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294650, 'name': CreateVM_Task, 'duration_secs': 0.407075} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.359890] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1177.360366] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.360533] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1177.360860] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1177.361321] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9ac586d-38f7-4be8-9665-d2f20b85ed1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.368906] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1177.368906] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5257fa3f-2baa-bce2-1f48-eab2bbdb6873" [ 1177.368906] env[62814]: _type = "Task" [ 1177.368906] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.377159] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5257fa3f-2baa-bce2-1f48-eab2bbdb6873, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.397430] env[62814]: WARNING nova.network.neutron [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] 1de03cb0-69fa-4200-8fa1-26375f44369a already exists in list: networks containing: ['1de03cb0-69fa-4200-8fa1-26375f44369a']. ignoring it [ 1177.460151] env[62814]: DEBUG nova.compute.utils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1177.466329] env[62814]: DEBUG nova.compute.manager [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Received event network-changed-461b5db0-7363-4a49-8a69-75f620fd7861 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1177.466516] env[62814]: DEBUG nova.compute.manager [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Refreshing instance network info cache due to event network-changed-461b5db0-7363-4a49-8a69-75f620fd7861. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1177.466741] env[62814]: DEBUG oslo_concurrency.lockutils [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] Acquiring lock "refresh_cache-bf2ba498-7c47-4985-a84a-475037deec55" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.466890] env[62814]: DEBUG oslo_concurrency.lockutils [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] Acquired lock "refresh_cache-bf2ba498-7c47-4985-a84a-475037deec55" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1177.470303] env[62814]: DEBUG nova.network.neutron [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Refreshing network info cache for port 461b5db0-7363-4a49-8a69-75f620fd7861 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1177.485544] env[62814]: DEBUG nova.compute.manager [req-bc7a18c5-7d34-4465-817e-20d91d061b36 req-2088087b-9d8e-4264-b40e-e01a21475adc service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Received event network-vif-deleted-26a8d6bc-23fd-40f8-ac9c-cea10954ca94 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1177.485890] env[62814]: INFO nova.compute.manager [req-bc7a18c5-7d34-4465-817e-20d91d061b36 req-2088087b-9d8e-4264-b40e-e01a21475adc service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Neutron deleted interface 26a8d6bc-23fd-40f8-ac9c-cea10954ca94; detaching it from the instance and deleting it from the info cache [ 1177.487271] env[62814]: DEBUG nova.network.neutron [req-bc7a18c5-7d34-4465-817e-20d91d061b36 req-2088087b-9d8e-4264-b40e-e01a21475adc service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.694019] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Releasing lock "refresh_cache-53e13866-0174-4867-9fd9-2ea3f5ee399d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.694019] env[62814]: DEBUG nova.compute.manager [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1177.694019] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1177.694019] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976b4ce6-0366-429d-b770-befe08513192 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.710058] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1177.710508] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7eb3b957-9704-4f76-9c96-b33039fa309d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.718287] env[62814]: DEBUG oslo_vmware.api [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1177.718287] env[62814]: value = "task-4294652" [ 1177.718287] env[62814]: _type = "Task" [ 1177.718287] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.725216] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 316931ae-7a62-4bac-81e4-1fee9a36164e] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1177.733098] env[62814]: DEBUG oslo_vmware.api [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294652, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.833235] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294651, 'name': Rename_Task, 'duration_secs': 0.255244} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.833601] env[62814]: DEBUG nova.network.neutron [-] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.835196] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1177.835491] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be16ffd2-9efd-4477-ac6b-57f4c53e1e61 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.843596] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1177.843596] env[62814]: value = "task-4294653" [ 1177.843596] env[62814]: _type = "Task" [ 1177.843596] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.855647] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294653, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.858213] env[62814]: DEBUG nova.network.neutron [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [{"id": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "address": "fa:16:3e:e8:10:0d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c6b72d7-d1", "ovs_interfaceid": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "88931aee-29b6-4f58-9ee6-0210612c8211", "address": "fa:16:3e:49:23:fd", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88931aee-29", "ovs_interfaceid": "88931aee-29b6-4f58-9ee6-0210612c8211", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.880814] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5257fa3f-2baa-bce2-1f48-eab2bbdb6873, 'name': SearchDatastore_Task, 'duration_secs': 0.00907} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1177.880814] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1177.880814] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1177.881119] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1177.881119] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1177.881236] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1177.881515] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-028f861a-0c9f-4974-9690-8390f6f9483e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.894140] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1177.894480] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1177.895814] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5abae257-f8b0-42d2-a90f-66d462049b4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.905898] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1177.905898] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c8e268-af1a-874d-0da2-8e77b0884010" [ 1177.905898] env[62814]: _type = "Task" [ 1177.905898] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.914840] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c8e268-af1a-874d-0da2-8e77b0884010, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.962959] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ea9e6f-eb56-4c89-adee-439d73fabb74 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.969056] env[62814]: INFO nova.virt.block_device [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Booting with volume 3e53f486-225a-44b7-bf8a-b91c6ac86fda at /dev/sdb [ 1177.982635] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a839b600-fda2-4275-a726-2d48d07ee308 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.016433] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f998b8bd-0b2d-4450-b87c-0038825b364b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.018936] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace51639-efd1-46c6-b084-75dc1211349e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.023545] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa25ccd2-89c5-4627-ba04-f5eca0df3224 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.033052] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0cfca9-db5a-4209-add4-02c68c388486 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.037258] env[62814]: DEBUG nova.network.neutron [-] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.042848] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbdda681-8694-4537-b3a9-163252bfb623 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.056214] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fafdffbb-e3d1-43c7-9bc1-0172352a23a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.079912] env[62814]: DEBUG nova.compute.provider_tree [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1178.103533] env[62814]: DEBUG nova.compute.manager [req-bc7a18c5-7d34-4465-817e-20d91d061b36 req-2088087b-9d8e-4264-b40e-e01a21475adc service nova] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Detach interface failed, port_id=26a8d6bc-23fd-40f8-ac9c-cea10954ca94, reason: Instance 881e9481-a5a9-489b-8abd-8efdaa239a11 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1178.104744] env[62814]: DEBUG nova.scheduler.client.report [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1178.109735] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-266452a2-ba9f-42b4-9857-1603dd80aa2c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.118951] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b62c56dc-8e61-4b07-a739-c0feedd11071 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.151200] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ef3c11-4d79-4b0d-84c8-64f4fb048a94 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.157771] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88aa92f-caa6-47d2-92bf-71530c496825 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.171335] env[62814]: DEBUG nova.virt.block_device [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating existing volume attachment record: 5d72190a-51a7-4c35-afae-c61d93fc3655 {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1178.228947] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 4ebc3280-64a7-4e9b-8385-2128357d3422] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1178.230967] env[62814]: DEBUG oslo_vmware.api [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294652, 'name': PowerOffVM_Task, 'duration_secs': 0.121033} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.231644] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1178.231937] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1178.232094] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eba70f08-5873-499f-803c-bd87d819bc16 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.234266] env[62814]: DEBUG nova.network.neutron [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Updated VIF entry in instance network info cache for port 461b5db0-7363-4a49-8a69-75f620fd7861. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1178.234631] env[62814]: DEBUG nova.network.neutron [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Updating instance_info_cache with network_info: [{"id": "461b5db0-7363-4a49-8a69-75f620fd7861", "address": "fa:16:3e:23:01:29", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap461b5db0-73", "ovs_interfaceid": "461b5db0-7363-4a49-8a69-75f620fd7861", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.256827] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1178.257106] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1178.257292] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Deleting the datastore file [datastore2] 53e13866-0174-4867-9fd9-2ea3f5ee399d {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1178.258370] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-93c210b8-1f5f-4ff4-934e-f3ad2298adae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.265255] env[62814]: DEBUG oslo_vmware.api [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for the task: (returnval){ [ 1178.265255] env[62814]: value = "task-4294655" [ 1178.265255] env[62814]: _type = "Task" [ 1178.265255] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.276123] env[62814]: DEBUG oslo_vmware.api [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294655, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.338742] env[62814]: INFO nova.compute.manager [-] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Took 1.29 seconds to deallocate network for instance. [ 1178.353305] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294653, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.361137] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1178.361857] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.362065] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1178.363407] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a36f09-c6ce-4098-8f46-404dcf3f5462 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.380576] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1178.380862] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1178.381048] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1178.381264] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1178.381624] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1178.381624] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1178.381909] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1178.382257] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1178.382257] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1178.382429] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1178.382542] env[62814]: DEBUG nova.virt.hardware [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1178.390343] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Reconfiguring VM to attach interface {{(pid=62814) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1178.391468] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-106ad8f3-6d2a-43a9-b369-19d8420e29be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.411804] env[62814]: DEBUG oslo_vmware.api [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1178.411804] env[62814]: value = "task-4294659" [ 1178.411804] env[62814]: _type = "Task" [ 1178.411804] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.419016] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c8e268-af1a-874d-0da2-8e77b0884010, 'name': SearchDatastore_Task, 'duration_secs': 0.009617} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.420470] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f786d4c1-d28f-4856-a00b-0234b8013446 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.425699] env[62814]: DEBUG oslo_vmware.api [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294659, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.428769] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1178.428769] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd240d-9920-d31a-fb29-68ce4062375c" [ 1178.428769] env[62814]: _type = "Task" [ 1178.428769] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.436523] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd240d-9920-d31a-fb29-68ce4062375c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.541114] env[62814]: INFO nova.compute.manager [-] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Took 1.64 seconds to deallocate network for instance. [ 1178.594237] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.594608] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.613530] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.394s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1178.632562] env[62814]: INFO nova.scheduler.client.report [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Deleted allocations for instance 29468dc9-4254-4ac6-989c-c7fa0c4cc682 [ 1178.682265] env[62814]: DEBUG nova.compute.manager [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Stashing vm_state: active {{(pid=62814) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1178.732523] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: e727eeb1-c5d0-4591-80bb-31746bf976a2] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1178.737397] env[62814]: DEBUG oslo_concurrency.lockutils [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] Releasing lock "refresh_cache-bf2ba498-7c47-4985-a84a-475037deec55" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1178.737748] env[62814]: DEBUG nova.compute.manager [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received event network-vif-plugged-88931aee-29b6-4f58-9ee6-0210612c8211 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1178.737989] env[62814]: DEBUG oslo_concurrency.lockutils [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.738297] env[62814]: DEBUG oslo_concurrency.lockutils [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.738546] env[62814]: DEBUG oslo_concurrency.lockutils [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1178.738782] env[62814]: DEBUG nova.compute.manager [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] No waiting events found dispatching network-vif-plugged-88931aee-29b6-4f58-9ee6-0210612c8211 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1178.738980] env[62814]: WARNING nova.compute.manager [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received unexpected event network-vif-plugged-88931aee-29b6-4f58-9ee6-0210612c8211 for instance with vm_state active and task_state None. [ 1178.739162] env[62814]: DEBUG nova.compute.manager [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received event network-changed-88931aee-29b6-4f58-9ee6-0210612c8211 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1178.739363] env[62814]: DEBUG nova.compute.manager [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Refreshing instance network info cache due to event network-changed-88931aee-29b6-4f58-9ee6-0210612c8211. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1178.739555] env[62814]: DEBUG oslo_concurrency.lockutils [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] Acquiring lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.739709] env[62814]: DEBUG oslo_concurrency.lockutils [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] Acquired lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1178.739899] env[62814]: DEBUG nova.network.neutron [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Refreshing network info cache for port 88931aee-29b6-4f58-9ee6-0210612c8211 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1178.775856] env[62814]: DEBUG oslo_vmware.api [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Task: {'id': task-4294655, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.088698} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.776141] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1178.776343] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1178.776526] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1178.776698] env[62814]: INFO nova.compute.manager [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Took 1.08 seconds to destroy the instance on the hypervisor. [ 1178.776935] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1178.777143] env[62814]: DEBUG nova.compute.manager [-] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1178.777238] env[62814]: DEBUG nova.network.neutron [-] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1178.798237] env[62814]: DEBUG nova.network.neutron [-] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1178.848352] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1178.848630] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1178.848823] env[62814]: DEBUG nova.objects.instance [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lazy-loading 'resources' on Instance uuid 881e9481-a5a9-489b-8abd-8efdaa239a11 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1178.855099] env[62814]: DEBUG oslo_vmware.api [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294653, 'name': PowerOnVM_Task, 'duration_secs': 0.749761} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.855410] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1178.922163] env[62814]: DEBUG oslo_vmware.api [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.939176] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd240d-9920-d31a-fb29-68ce4062375c, 'name': SearchDatastore_Task, 'duration_secs': 0.027539} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.939447] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1178.939666] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] bf2ba498-7c47-4985-a84a-475037deec55/bf2ba498-7c47-4985-a84a-475037deec55.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1178.939931] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1f87710f-9eda-4edb-ab72-9161ddb340fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.946802] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1178.946802] env[62814]: value = "task-4294660" [ 1178.946802] env[62814]: _type = "Task" [ 1178.946802] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.955875] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294660, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.972701] env[62814]: DEBUG nova.compute.manager [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1178.973764] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0029022e-5ea9-4a45-b9fb-223368a05810 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.049018] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1179.098198] env[62814]: DEBUG nova.compute.utils [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1179.140406] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e1281229-4fa4-4c84-a94d-e8ed7d095b8f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "29468dc9-4254-4ac6-989c-c7fa0c4cc682" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.705s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.202146] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1179.236190] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: a68901a9-9a9b-4127-bca9-64f98dfb151f] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1179.300818] env[62814]: DEBUG nova.network.neutron [-] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.427019] env[62814]: DEBUG oslo_vmware.api [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294659, 'name': ReconfigVM_Task, 'duration_secs': 0.599294} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.427558] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1179.427780] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Reconfigured VM to attach interface {{(pid=62814) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1179.462464] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294660, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.491275] env[62814]: DEBUG nova.network.neutron [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updated VIF entry in instance network info cache for port 88931aee-29b6-4f58-9ee6-0210612c8211. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1179.491733] env[62814]: DEBUG nova.network.neutron [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [{"id": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "address": "fa:16:3e:e8:10:0d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c6b72d7-d1", "ovs_interfaceid": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "88931aee-29b6-4f58-9ee6-0210612c8211", "address": "fa:16:3e:49:23:fd", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88931aee-29", "ovs_interfaceid": "88931aee-29b6-4f58-9ee6-0210612c8211", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1179.497495] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00fd2dd7-d0a3-4d9e-bbad-7600161f9eab tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 23.077s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.568342] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a4254e-0f04-4fa7-951a-4fc8b7c5731e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.576755] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56115e21-0d06-439b-88c9-6513fc6433bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.607847] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.013s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.609135] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-274f440b-f5bc-4823-9d7e-85ef66570e5a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.616862] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd14f168-99da-42e7-9628-d45f61d2dd50 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.630854] env[62814]: DEBUG nova.compute.provider_tree [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1179.739756] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 885e0798-2fa6-4f6b-82be-517b6d1168d2] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1179.804574] env[62814]: INFO nova.compute.manager [-] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Took 1.03 seconds to deallocate network for instance. [ 1179.932982] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f7a62be1-3958-4913-ad60-ddb5ba76301a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.104s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1179.958287] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294660, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.624834} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.959123] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] bf2ba498-7c47-4985-a84a-475037deec55/bf2ba498-7c47-4985-a84a-475037deec55.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1179.959341] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1179.959588] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10a0f724-1266-4361-aac2-e8521bb57968 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.965539] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1179.965539] env[62814]: value = "task-4294661" [ 1179.965539] env[62814]: _type = "Task" [ 1179.965539] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.975139] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294661, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.998533] env[62814]: DEBUG oslo_concurrency.lockutils [req-c3b92893-8a01-42c9-ac88-c374fe0b5ec9 req-887f0931-cf8b-427a-90c9-2c01861bf36a service nova] Releasing lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1180.134452] env[62814]: DEBUG nova.scheduler.client.report [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1180.243632] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: fd1867e3-2523-4969-a157-b14c650f3779] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1180.310417] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1180.474391] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294661, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.221871} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.474707] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1180.475553] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019ab7d5-edf1-4a93-ba97-7a0f8da12bd1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.498770] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] bf2ba498-7c47-4985-a84a-475037deec55/bf2ba498-7c47-4985-a84a-475037deec55.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1180.499056] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7a54193d-8e60-4830-937c-b065d463cd3b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.518282] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1180.518282] env[62814]: value = "task-4294663" [ 1180.518282] env[62814]: _type = "Task" [ 1180.518282] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.526900] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294663, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.640202] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1180.642640] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.594s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1180.642875] env[62814]: DEBUG nova.objects.instance [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lazy-loading 'resources' on Instance uuid f00d414e-1dfd-43cf-9245-7cbeea8850b3 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1180.662047] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1180.662316] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1180.662567] env[62814]: INFO nova.compute.manager [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Attaching volume 808c711a-a0ee-42d1-96d6-cb028b2a9fe7 to /dev/sdb [ 1180.665195] env[62814]: INFO nova.scheduler.client.report [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted allocations for instance 881e9481-a5a9-489b-8abd-8efdaa239a11 [ 1180.701392] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0296d4a9-4985-4e50-b3ed-2c6595a0b32f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.708296] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b6171f-11dd-4198-880d-f91d30b9b669 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.722887] env[62814]: DEBUG nova.virt.block_device [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating existing volume attachment record: 2f0b9459-08e6-4f78-85cb-8b6a5a5b7d6a {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1180.746979] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: b5ddd05f-2027-4edf-84c5-0d2f537a95e4] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1180.811039] env[62814]: DEBUG nova.compute.manager [req-16c8aa33-c969-4d75-bcff-45f4552877b8 req-18277671-586c-4a53-8a94-5f8f87692573 service nova] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Received event network-vif-deleted-3d97fdd4-3a3a-48c2-8729-5e78d0f6a8d5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1181.028122] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294663, 'name': ReconfigVM_Task, 'duration_secs': 0.292082} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.028122] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Reconfigured VM instance instance-0000006f to attach disk [datastore2] bf2ba498-7c47-4985-a84a-475037deec55/bf2ba498-7c47-4985-a84a-475037deec55.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1181.028645] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49e93a82-6333-4c5c-911d-a8d5910551b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.036267] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1181.036267] env[62814]: value = "task-4294665" [ 1181.036267] env[62814]: _type = "Task" [ 1181.036267] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.045134] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294665, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.173506] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7585a43b-d235-4018-8445-f19c2969bc50 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "881e9481-a5a9-489b-8abd-8efdaa239a11" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.245s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1181.254995] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 40e2d845-0211-4c84-aef7-94014f999e1d] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1181.358890] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c551ee-acf1-488d-9a96-d8320a03aa39 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.367188] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72e89106-fd80-4464-87ee-194fc05d6ba4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.402689] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87442a3c-323f-4387-8c06-45589ecd7f17 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.411431] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04eccebc-9f21-4f86-a03d-03860d045aa4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.427728] env[62814]: DEBUG nova.compute.provider_tree [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1181.484160] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "d65ccde2-b31a-4032-8795-1d609fdfcc73" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1181.484475] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "d65ccde2-b31a-4032-8795-1d609fdfcc73" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1181.484710] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "d65ccde2-b31a-4032-8795-1d609fdfcc73-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1181.484939] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "d65ccde2-b31a-4032-8795-1d609fdfcc73-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1181.485160] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "d65ccde2-b31a-4032-8795-1d609fdfcc73-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1181.487544] env[62814]: INFO nova.compute.manager [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Terminating instance [ 1181.545653] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294665, 'name': Rename_Task, 'duration_secs': 0.293029} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.545861] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1181.546138] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93634640-df5d-4c03-a303-d9a5006d45a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.556623] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1181.556623] env[62814]: value = "task-4294668" [ 1181.556623] env[62814]: _type = "Task" [ 1181.556623] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.561832] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294668, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.755058] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 267c1ef2-ccc3-470e-943d-5fe9203c4dbc] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1181.931625] env[62814]: DEBUG nova.scheduler.client.report [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1181.996337] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "refresh_cache-d65ccde2-b31a-4032-8795-1d609fdfcc73" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.997403] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquired lock "refresh_cache-d65ccde2-b31a-4032-8795-1d609fdfcc73" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1181.997403] env[62814]: DEBUG nova.network.neutron [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1182.063703] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294668, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.072668] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-fec67180-c78f-4b6c-b106-a0d1ab127ec4" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.072910] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-fec67180-c78f-4b6c-b106-a0d1ab127ec4" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.073348] env[62814]: DEBUG nova.objects.instance [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'flavor' on Instance uuid 2d7a18f3-456b-470c-a759-632e5d79a8f0 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1182.258036] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 22429810-3ea7-4472-9a90-018d6efcba1d] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1182.437805] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.795s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1182.440642] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.239s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.468096] env[62814]: INFO nova.scheduler.client.report [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleted allocations for instance f00d414e-1dfd-43cf-9245-7cbeea8850b3 [ 1182.480808] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3c7ada-5d69-4bb9-b86c-15c9df4c1f2d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.489182] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-61043b69-6014-45bc-b6e7-879836c12908 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Suspending the VM {{(pid=62814) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1182.489182] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-8b520051-d8bf-4866-9f0b-f767de6641af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.496569] env[62814]: DEBUG oslo_vmware.api [None req-61043b69-6014-45bc-b6e7-879836c12908 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1182.496569] env[62814]: value = "task-4294669" [ 1182.496569] env[62814]: _type = "Task" [ 1182.496569] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.509997] env[62814]: DEBUG oslo_vmware.api [None req-61043b69-6014-45bc-b6e7-879836c12908 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294669, 'name': SuspendVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.529765] env[62814]: DEBUG nova.network.neutron [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1182.564771] env[62814]: DEBUG oslo_vmware.api [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294668, 'name': PowerOnVM_Task, 'duration_secs': 0.636067} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.565174] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1182.565377] env[62814]: INFO nova.compute.manager [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Took 9.07 seconds to spawn the instance on the hypervisor. [ 1182.565647] env[62814]: DEBUG nova.compute.manager [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1182.566681] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3475cc18-bddb-4a1b-a168-bd46ea74aae2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.645167] env[62814]: DEBUG nova.network.neutron [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.757142] env[62814]: DEBUG nova.objects.instance [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'pci_requests' on Instance uuid 2d7a18f3-456b-470c-a759-632e5d79a8f0 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1182.761653] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 508321ab-ce10-4953-a9e3-193b9975bec7] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1182.915979] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1182.917964] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.002s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1182.947710] env[62814]: INFO nova.compute.claims [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1182.977968] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf94f8c0-190f-4019-975f-ec14f98a5753 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "f00d414e-1dfd-43cf-9245-7cbeea8850b3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.815s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1183.008357] env[62814]: DEBUG oslo_vmware.api [None req-61043b69-6014-45bc-b6e7-879836c12908 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294669, 'name': SuspendVM_Task} progress is 62%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.091903] env[62814]: INFO nova.compute.manager [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Took 14.51 seconds to build instance. [ 1183.093508] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.093821] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1183.148036] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Releasing lock "refresh_cache-d65ccde2-b31a-4032-8795-1d609fdfcc73" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1183.148411] env[62814]: DEBUG nova.compute.manager [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1183.148658] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1183.149490] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c25faeb-ccb9-4d1a-a420-ab8f3b840a5a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.156895] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1183.157143] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-243e72bd-704a-4371-96d6-d672ba1f1b0b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.163296] env[62814]: DEBUG oslo_vmware.api [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1183.163296] env[62814]: value = "task-4294670" [ 1183.163296] env[62814]: _type = "Task" [ 1183.163296] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.171668] env[62814]: DEBUG oslo_vmware.api [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294670, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.262836] env[62814]: DEBUG nova.objects.base [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Object Instance<2d7a18f3-456b-470c-a759-632e5d79a8f0> lazy-loaded attributes: flavor,pci_requests {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1183.263147] env[62814]: DEBUG nova.network.neutron [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1183.265672] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: f54b8bb9-69b7-4bb4-a82c-9f796050e719] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1183.328856] env[62814]: DEBUG nova.policy [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5779eaa60ec44b2b80660e521b598eb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10018060297c409d82935ad626cc54c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1183.420998] env[62814]: DEBUG nova.compute.utils [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1183.454426] env[62814]: INFO nova.compute.resource_tracker [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating resource usage from migration c7cb01ac-d4df-41dd-b19a-6835a05449a3 [ 1183.509973] env[62814]: DEBUG oslo_vmware.api [None req-61043b69-6014-45bc-b6e7-879836c12908 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294669, 'name': SuspendVM_Task, 'duration_secs': 0.967837} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.510766] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-61043b69-6014-45bc-b6e7-879836c12908 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Suspended the VM {{(pid=62814) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1183.510766] env[62814]: DEBUG nova.compute.manager [None req-61043b69-6014-45bc-b6e7-879836c12908 tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1183.512184] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d948b8-3c7e-4740-831f-bcf500ada1d8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.593070] env[62814]: DEBUG oslo_concurrency.lockutils [None req-11757dd7-2756-4ef4-8b3a-5f3d6d856e70 tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "bf2ba498-7c47-4985-a84a-475037deec55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.025s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1183.595575] env[62814]: DEBUG nova.compute.manager [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1183.656766] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837ef33d-d134-46fc-812b-9871506180f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.664276] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-217c96d3-fe1b-44fc-a52d-61db79cfe48a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.675503] env[62814]: DEBUG oslo_vmware.api [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294670, 'name': PowerOffVM_Task, 'duration_secs': 0.243587} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.699357] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1183.699569] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1183.699998] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80371583-0f24-43c7-8b93-ec7f53c10caa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.702060] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3697329-a8f8-40d8-954b-3a7bfa7bcaa7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.709570] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b92180-1dfb-4ceb-94eb-ee2c3bf54f17 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.722650] env[62814]: DEBUG nova.compute.provider_tree [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1183.727154] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1183.727361] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1183.727540] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Deleting the datastore file [datastore2] d65ccde2-b31a-4032-8795-1d609fdfcc73 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1183.727778] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-83d68eec-32a7-43ea-a634-d399cd425930 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.734810] env[62814]: DEBUG oslo_vmware.api [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for the task: (returnval){ [ 1183.734810] env[62814]: value = "task-4294673" [ 1183.734810] env[62814]: _type = "Task" [ 1183.734810] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.742814] env[62814]: DEBUG oslo_vmware.api [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.769773] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 5c574786-64a4-4f07-a267-101ecaaa6938] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1183.785991] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1183.924071] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1184.113217] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1184.226536] env[62814]: DEBUG nova.scheduler.client.report [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1184.246388] env[62814]: DEBUG oslo_vmware.api [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294673, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.274471] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 0fa2c6ca-1cc7-4b48-983e-fc1e18087f0a] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1184.732038] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.291s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1184.732306] env[62814]: INFO nova.compute.manager [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Migrating [ 1184.740630] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.430s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1184.740933] env[62814]: DEBUG nova.objects.instance [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lazy-loading 'resources' on Instance uuid 53e13866-0174-4867-9fd9-2ea3f5ee399d {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1184.765491] env[62814]: DEBUG oslo_vmware.api [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Task: {'id': task-4294673, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.649288} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.765818] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1184.766075] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1184.766295] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1184.766661] env[62814]: INFO nova.compute.manager [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1184.766790] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1184.767120] env[62814]: DEBUG nova.compute.manager [-] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1184.767184] env[62814]: DEBUG nova.network.neutron [-] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1184.777529] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 99864c68-5f11-4119-b8e3-3aa8719f267b] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1184.791674] env[62814]: DEBUG nova.network.neutron [-] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1184.961037] env[62814]: DEBUG nova.network.neutron [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Successfully updated port: fec67180-c78f-4b6c-b106-a0d1ab127ec4 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1184.968863] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98d7a4a7-7bb8-417e-adea-e0c2f3348b8e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.978232] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e115c024-ba2c-40fd-a0d6-cc3c55c7391a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.011597] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1185.011888] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1185.012163] env[62814]: INFO nova.compute.manager [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Attaching volume c2969fdc-ffa5-4fde-b88c-26966e10e336 to /dev/sdb [ 1185.015336] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8efc662-f6c7-4d41-b1fa-11254d8f87e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.027094] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841a9ba2-6430-479e-b37d-88e52d057750 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.041201] env[62814]: DEBUG nova.compute.provider_tree [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1185.046025] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd0aa94-96d9-4723-bb6d-d75824ad69c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.052876] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac5ad020-1810-43e1-a2b8-56c4e1aa94b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.066701] env[62814]: DEBUG nova.virt.block_device [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Updating existing volume attachment record: b57c2613-e063-4c9a-9da4-a0633693b9be {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1185.248611] env[62814]: DEBUG nova.compute.manager [req-04577a15-aeaa-45ed-8550-e809e1f01004 req-41840089-cb97-4641-8d5b-c178db296a13 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received event network-vif-plugged-fec67180-c78f-4b6c-b106-a0d1ab127ec4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1185.248873] env[62814]: DEBUG oslo_concurrency.lockutils [req-04577a15-aeaa-45ed-8550-e809e1f01004 req-41840089-cb97-4641-8d5b-c178db296a13 service nova] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1185.249241] env[62814]: DEBUG oslo_concurrency.lockutils [req-04577a15-aeaa-45ed-8550-e809e1f01004 req-41840089-cb97-4641-8d5b-c178db296a13 service nova] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1185.249643] env[62814]: DEBUG oslo_concurrency.lockutils [req-04577a15-aeaa-45ed-8550-e809e1f01004 req-41840089-cb97-4641-8d5b-c178db296a13 service nova] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1185.249643] env[62814]: DEBUG nova.compute.manager [req-04577a15-aeaa-45ed-8550-e809e1f01004 req-41840089-cb97-4641-8d5b-c178db296a13 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] No waiting events found dispatching network-vif-plugged-fec67180-c78f-4b6c-b106-a0d1ab127ec4 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1185.249802] env[62814]: WARNING nova.compute.manager [req-04577a15-aeaa-45ed-8550-e809e1f01004 req-41840089-cb97-4641-8d5b-c178db296a13 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received unexpected event network-vif-plugged-fec67180-c78f-4b6c-b106-a0d1ab127ec4 for instance with vm_state active and task_state None. [ 1185.259466] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.259599] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1185.259777] env[62814]: DEBUG nova.network.neutron [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1185.280720] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: af6ef867-dc9a-4db6-8582-a6198e5caa77] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1185.294962] env[62814]: DEBUG nova.network.neutron [-] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.464676] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1185.464676] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1185.464676] env[62814]: DEBUG nova.network.neutron [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1185.544669] env[62814]: DEBUG nova.scheduler.client.report [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1185.775561] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1185.775814] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845857', 'volume_id': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'name': 'volume-808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b1d88997-e52f-41bd-b1b4-dd096d20d60a', 'attached_at': '', 'detached_at': '', 'volume_id': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'serial': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1185.776759] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2197c11-11ad-4563-ac3f-c85ebb4c26a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.796508] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: d7332882-19b3-4ab3-8ea6-51d33c584844] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1185.799391] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c677e7e-12c7-4cae-999f-589d1f714e99 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.802824] env[62814]: INFO nova.compute.manager [-] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Took 1.04 seconds to deallocate network for instance. [ 1185.836217] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] volume-808c711a-a0ee-42d1-96d6-cb028b2a9fe7/volume-808c711a-a0ee-42d1-96d6-cb028b2a9fe7.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1185.837103] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63418663-92ff-43a0-854e-3cefe6568bff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.859752] env[62814]: DEBUG oslo_vmware.api [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1185.859752] env[62814]: value = "task-4294675" [ 1185.859752] env[62814]: _type = "Task" [ 1185.859752] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.871169] env[62814]: DEBUG oslo_vmware.api [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294675, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.994696] env[62814]: INFO nova.compute.manager [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Resuming [ 1185.995454] env[62814]: DEBUG nova.objects.instance [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lazy-loading 'flavor' on Instance uuid edcdb1e5-ed9d-49e6-97e6-ea7629682547 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1186.021844] env[62814]: WARNING nova.network.neutron [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] 1de03cb0-69fa-4200-8fa1-26375f44369a already exists in list: networks containing: ['1de03cb0-69fa-4200-8fa1-26375f44369a']. ignoring it [ 1186.022090] env[62814]: WARNING nova.network.neutron [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] 1de03cb0-69fa-4200-8fa1-26375f44369a already exists in list: networks containing: ['1de03cb0-69fa-4200-8fa1-26375f44369a']. ignoring it [ 1186.049094] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.308s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.052259] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.266s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.052259] env[62814]: DEBUG nova.objects.instance [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'pci_requests' on Instance uuid 1f13051e-e51b-4981-9445-d5420c3c1818 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1186.074922] env[62814]: INFO nova.scheduler.client.report [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Deleted allocations for instance 53e13866-0174-4867-9fd9-2ea3f5ee399d [ 1186.083142] env[62814]: DEBUG nova.network.neutron [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance_info_cache with network_info: [{"id": "580abcb7-b7f2-437a-8a84-738e83af0751", "address": "fa:16:3e:f1:6c:e8", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap580abcb7-b7", "ovs_interfaceid": "580abcb7-b7f2-437a-8a84-738e83af0751", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.304236] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 98ba2786-023f-4d36-bdd0-e38bdd23d73c] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1186.338199] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.371273] env[62814]: DEBUG oslo_vmware.api [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294675, 'name': ReconfigVM_Task, 'duration_secs': 0.371823} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.371574] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfigured VM instance instance-00000067 to attach disk [datastore1] volume-808c711a-a0ee-42d1-96d6-cb028b2a9fe7/volume-808c711a-a0ee-42d1-96d6-cb028b2a9fe7.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1186.376285] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b649251e-38b7-4a9e-bb48-937e8b87bc55 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.391831] env[62814]: DEBUG oslo_vmware.api [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1186.391831] env[62814]: value = "task-4294676" [ 1186.391831] env[62814]: _type = "Task" [ 1186.391831] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.402454] env[62814]: DEBUG oslo_vmware.api [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294676, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.431137] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "bf2ba498-7c47-4985-a84a-475037deec55" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.431359] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "bf2ba498-7c47-4985-a84a-475037deec55" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.431746] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "bf2ba498-7c47-4985-a84a-475037deec55-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1186.431966] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "bf2ba498-7c47-4985-a84a-475037deec55-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1186.432173] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "bf2ba498-7c47-4985-a84a-475037deec55-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.434641] env[62814]: INFO nova.compute.manager [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Terminating instance [ 1186.556692] env[62814]: DEBUG nova.objects.instance [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'numa_topology' on Instance uuid 1f13051e-e51b-4981-9445-d5420c3c1818 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1186.585402] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bfadfc8b-77b5-40d5-97fa-3db7fc5441fe tempest-ServerShowV254Test-106885106 tempest-ServerShowV254Test-106885106-project-member] Lock "53e13866-0174-4867-9fd9-2ea3f5ee399d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.573s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1186.586552] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1186.729110] env[62814]: DEBUG nova.network.neutron [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [{"id": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "address": "fa:16:3e:e8:10:0d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c6b72d7-d1", "ovs_interfaceid": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "88931aee-29b6-4f58-9ee6-0210612c8211", "address": "fa:16:3e:49:23:fd", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88931aee-29", "ovs_interfaceid": "88931aee-29b6-4f58-9ee6-0210612c8211", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fec67180-c78f-4b6c-b106-a0d1ab127ec4", "address": "fa:16:3e:00:36:d2", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfec67180-c7", "ovs_interfaceid": "fec67180-c78f-4b6c-b106-a0d1ab127ec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1186.809075] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 310e22c3-ff0a-4e6b-aa46-8812fa014dfa] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1186.902707] env[62814]: DEBUG oslo_vmware.api [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294676, 'name': ReconfigVM_Task, 'duration_secs': 0.145302} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.903066] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845857', 'volume_id': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'name': 'volume-808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'b1d88997-e52f-41bd-b1b4-dd096d20d60a', 'attached_at': '', 'detached_at': '', 'volume_id': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'serial': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1186.937961] env[62814]: DEBUG nova.compute.manager [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1186.938211] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1186.939084] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6291f82-7527-41ae-9ebb-8c84ca280bac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.946546] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1186.946784] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-59418ddd-66f8-447a-87e5-c74a3f1b571e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.953827] env[62814]: DEBUG oslo_vmware.api [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1186.953827] env[62814]: value = "task-4294677" [ 1186.953827] env[62814]: _type = "Task" [ 1186.953827] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.960676] env[62814]: DEBUG oslo_vmware.api [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294677, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.059691] env[62814]: INFO nova.compute.claims [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1187.232465] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1187.233172] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.233352] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1187.234226] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b69ae24c-52fb-4011-a022-70761e67a158 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1187.256824] env[62814]: DEBUG nova.virt.hardware [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1187.261508] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Reconfiguring VM to attach interface {{(pid=62814) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1187.261827] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab0dd45a-1f3a-469d-ab9f-4fc373216a47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.279626] env[62814]: DEBUG oslo_vmware.api [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1187.279626] env[62814]: value = "task-4294679" [ 1187.279626] env[62814]: _type = "Task" [ 1187.279626] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.290430] env[62814]: DEBUG oslo_vmware.api [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294679, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.312200] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 98884571-8101-4ae9-b2e3-aeed97e3618f] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1187.463633] env[62814]: DEBUG oslo_vmware.api [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294677, 'name': PowerOffVM_Task, 'duration_secs': 0.197277} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.464029] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1187.464029] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1187.464182] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-622027f1-e223-4361-b990-5378a22b8d68 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.505653] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.505915] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquired lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1187.506091] env[62814]: DEBUG nova.network.neutron [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1187.528938] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1187.529181] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1187.529368] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleting the datastore file [datastore2] bf2ba498-7c47-4985-a84a-475037deec55 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1187.529664] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-be1cf631-7404-421e-9631-8519fbe5bb58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.535896] env[62814]: DEBUG oslo_vmware.api [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1187.535896] env[62814]: value = "task-4294681" [ 1187.535896] env[62814]: _type = "Task" [ 1187.535896] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.544556] env[62814]: DEBUG oslo_vmware.api [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294681, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.554153] env[62814]: DEBUG nova.compute.manager [req-c9dd9df3-3f3c-4a01-9532-d0e1fdc9e6a0 req-c92e6e2e-4b4b-446b-b868-eefc28226f01 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received event network-changed-fec67180-c78f-4b6c-b106-a0d1ab127ec4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1187.554933] env[62814]: DEBUG nova.compute.manager [req-c9dd9df3-3f3c-4a01-9532-d0e1fdc9e6a0 req-c92e6e2e-4b4b-446b-b868-eefc28226f01 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Refreshing instance network info cache due to event network-changed-fec67180-c78f-4b6c-b106-a0d1ab127ec4. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1187.555560] env[62814]: DEBUG oslo_concurrency.lockutils [req-c9dd9df3-3f3c-4a01-9532-d0e1fdc9e6a0 req-c92e6e2e-4b4b-446b-b868-eefc28226f01 service nova] Acquiring lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.555560] env[62814]: DEBUG oslo_concurrency.lockutils [req-c9dd9df3-3f3c-4a01-9532-d0e1fdc9e6a0 req-c92e6e2e-4b4b-446b-b868-eefc28226f01 service nova] Acquired lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1187.555560] env[62814]: DEBUG nova.network.neutron [req-c9dd9df3-3f3c-4a01-9532-d0e1fdc9e6a0 req-c92e6e2e-4b4b-446b-b868-eefc28226f01 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Refreshing network info cache for port fec67180-c78f-4b6c-b106-a0d1ab127ec4 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1187.788418] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1187.788637] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1187.795294] env[62814]: DEBUG oslo_vmware.api [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294679, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.815621] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 1e259ec6-d31a-453b-87e0-baa446665d56] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1187.946539] env[62814]: DEBUG nova.objects.instance [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'flavor' on Instance uuid b1d88997-e52f-41bd-b1b4-dd096d20d60a {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1188.049493] env[62814]: DEBUG oslo_vmware.api [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294681, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171603} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.052016] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1188.052910] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1188.052910] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1188.052910] env[62814]: INFO nova.compute.manager [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1188.052910] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1188.053242] env[62814]: DEBUG nova.compute.manager [-] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1188.053242] env[62814]: DEBUG nova.network.neutron [-] [instance: bf2ba498-7c47-4985-a84a-475037deec55] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1188.106149] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2599207e-ac56-498f-bda4-c93a641cd11d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.127266] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance '40e3c00b-2129-476e-bc67-a0ef13781990' progress to 0 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1188.287823] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed46766-6f71-4514-9762-212c5614e099 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.293376] env[62814]: DEBUG nova.network.neutron [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updating instance_info_cache with network_info: [{"id": "9950de47-d55a-480b-b88e-818e09b9bc80", "address": "fa:16:3e:0c:71:2f", "network": {"id": "9cb04bbd-b270-4f5c-8416-b2354172a53a", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-896647291-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "869db97f30f544f49e6a0c0e54bea463", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ad4fcde7-8926-402a-a9b7-4878d2bc1cf6", "external-id": "nsx-vlan-transportzone-840", "segmentation_id": 840, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9950de47-d5", "ovs_interfaceid": "9950de47-d55a-480b-b88e-818e09b9bc80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.298919] env[62814]: DEBUG nova.compute.manager [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1188.302678] env[62814]: DEBUG oslo_vmware.api [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294679, 'name': ReconfigVM_Task, 'duration_secs': 0.590183} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.303479] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1188.303732] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Reconfigured VM to attach interface {{(pid=62814) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1188.309995] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9050d08-1ded-4e11-bd5c-c0d7e3298d46 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.340102] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 844737d8-d852-44bb-bf9d-e673e737ef33] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1188.343834] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e66338-6f9c-4532-add6-8470f2492eb5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.351374] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e97576-424f-4f02-8647-40d3b88b773c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.365297] env[62814]: DEBUG nova.compute.provider_tree [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1188.451894] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0d2dca28-0a1c-4cc0-9176-b9fd28ea6ee2 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.789s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1188.637979] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1188.638330] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5c8f0dcf-1c72-4333-95ae-0fb45918f073 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.649837] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1188.649837] env[62814]: value = "task-4294682" [ 1188.649837] env[62814]: _type = "Task" [ 1188.649837] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.662374] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294682, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.716060] env[62814]: DEBUG nova.network.neutron [req-c9dd9df3-3f3c-4a01-9532-d0e1fdc9e6a0 req-c92e6e2e-4b4b-446b-b868-eefc28226f01 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updated VIF entry in instance network info cache for port fec67180-c78f-4b6c-b106-a0d1ab127ec4. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1188.716660] env[62814]: DEBUG nova.network.neutron [req-c9dd9df3-3f3c-4a01-9532-d0e1fdc9e6a0 req-c92e6e2e-4b4b-446b-b868-eefc28226f01 service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [{"id": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "address": "fa:16:3e:e8:10:0d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c6b72d7-d1", "ovs_interfaceid": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "88931aee-29b6-4f58-9ee6-0210612c8211", "address": "fa:16:3e:49:23:fd", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88931aee-29", "ovs_interfaceid": "88931aee-29b6-4f58-9ee6-0210612c8211", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fec67180-c78f-4b6c-b106-a0d1ab127ec4", "address": "fa:16:3e:00:36:d2", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfec67180-c7", "ovs_interfaceid": "fec67180-c78f-4b6c-b106-a0d1ab127ec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.798855] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Releasing lock "refresh_cache-edcdb1e5-ed9d-49e6-97e6-ea7629682547" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1188.800209] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d08ac60-879e-4ca0-a4ff-d1473797374a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.813808] env[62814]: DEBUG oslo_concurrency.lockutils [None req-cfc62739-5914-46a8-9960-32de4be97df9 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-fec67180-c78f-4b6c-b106-a0d1ab127ec4" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.741s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1188.814889] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Resuming the VM {{(pid=62814) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1188.815164] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fae58bd2-8111-47e1-bb4e-198cf4595fa2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.821861] env[62814]: DEBUG oslo_vmware.api [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1188.821861] env[62814]: value = "task-4294683" [ 1188.821861] env[62814]: _type = "Task" [ 1188.821861] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1188.825796] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1188.831216] env[62814]: DEBUG oslo_vmware.api [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294683, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.843754] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 9be62576-5a05-473f-befd-b33f5fde9185] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1188.870540] env[62814]: DEBUG nova.scheduler.client.report [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1189.160987] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294682, 'name': PowerOffVM_Task, 'duration_secs': 0.202256} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.161292] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1189.161521] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance '40e3c00b-2129-476e-bc67-a0ef13781990' progress to 17 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1189.170621] env[62814]: DEBUG nova.network.neutron [-] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1189.220090] env[62814]: DEBUG oslo_concurrency.lockutils [req-c9dd9df3-3f3c-4a01-9532-d0e1fdc9e6a0 req-c92e6e2e-4b4b-446b-b868-eefc28226f01 service nova] Releasing lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1189.332366] env[62814]: DEBUG oslo_vmware.api [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294683, 'name': PowerOnVM_Task, 'duration_secs': 0.504863} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1189.332759] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Resumed the VM {{(pid=62814) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1189.332978] env[62814]: DEBUG nova.compute.manager [None req-ec623cf8-a3b2-48fc-ac1f-033eaad008dd tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1189.333926] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d0271d-13be-4fe2-bd58-65bf5cbeb3d2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.348063] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: cd89e4d5-c08e-42c9-aea8-84c20714e196] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1189.352176] env[62814]: DEBUG nova.compute.manager [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Stashing vm_state: active {{(pid=62814) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1189.375251] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.323s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1189.377559] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.264s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1189.379086] env[62814]: INFO nova.compute.claims [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1189.411578] env[62814]: INFO nova.network.neutron [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating port a7104ea3-596d-4d10-952c-ac861087e580 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1189.459918] env[62814]: DEBUG nova.compute.manager [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Stashing vm_state: active {{(pid=62814) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1189.585777] env[62814]: DEBUG nova.compute.manager [req-ff521fba-8152-4dc1-9df7-110532fefef7 req-99d60d2e-946a-49b8-ba28-6ab2a9be53fe service nova] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Received event network-vif-deleted-461b5db0-7363-4a49-8a69-75f620fd7861 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1189.610898] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1189.611220] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845858', 'volume_id': 'c2969fdc-ffa5-4fde-b88c-26966e10e336', 'name': 'volume-c2969fdc-ffa5-4fde-b88c-26966e10e336', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '33ab2d9e-e960-40ae-acf9-4fea0d9c830e', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2969fdc-ffa5-4fde-b88c-26966e10e336', 'serial': 'c2969fdc-ffa5-4fde-b88c-26966e10e336'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1189.612134] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5045e31-326d-4bce-9f41-fa02e4304770 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.629492] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5516a1b7-6f91-4b04-8c77-fd4eb7a7daca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.655275] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] volume-c2969fdc-ffa5-4fde-b88c-26966e10e336/volume-c2969fdc-ffa5-4fde-b88c-26966e10e336.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1189.655625] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cefe0309-f96c-492a-a826-084c3d18abc2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.670870] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1189.671126] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1189.671287] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1189.671470] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1189.671616] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1189.671760] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1189.672026] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1189.672228] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1189.672401] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1189.672567] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1189.672739] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1189.679163] env[62814]: INFO nova.compute.manager [-] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Took 1.63 seconds to deallocate network for instance. [ 1189.679163] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-99b98c78-0f11-4a0c-915e-f2d0f31d34c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.697435] env[62814]: DEBUG oslo_vmware.api [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1189.697435] env[62814]: value = "task-4294684" [ 1189.697435] env[62814]: _type = "Task" [ 1189.697435] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.699114] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1189.699114] env[62814]: value = "task-4294685" [ 1189.699114] env[62814]: _type = "Task" [ 1189.699114] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.710776] env[62814]: DEBUG oslo_vmware.api [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294684, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.714187] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294685, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1189.852497] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: a473cc7e-0f86-4a18-9789-0aabfff430ab] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1189.875609] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1189.977946] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1190.200423] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1190.212704] env[62814]: DEBUG oslo_vmware.api [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.215695] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294685, 'name': ReconfigVM_Task, 'duration_secs': 0.288193} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.216343] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance '40e3c00b-2129-476e-bc67-a0ef13781990' progress to 33 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1190.355789] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: ca09a49f-c263-45c6-a6f7-aa9fa6e8fd01] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1190.442964] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-88931aee-29b6-4f58-9ee6-0210612c8211" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1190.443182] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-88931aee-29b6-4f58-9ee6-0210612c8211" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1190.561028] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3a9b24-8817-45e9-9125-671b5b9a183b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.571219] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f818f6d-47f8-470b-9a2e-52f523c34c59 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.605985] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd91e2e-bc7e-4f6a-92d4-61a69f7eef58 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.613386] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5dcece-c98f-40d7-a4f2-17acc2a65b05 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.626793] env[62814]: DEBUG nova.compute.provider_tree [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.710111] env[62814]: DEBUG oslo_vmware.api [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294684, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.725522] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1190.725776] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1190.725938] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1190.726136] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1190.726283] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1190.726431] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1190.726636] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1190.726815] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1190.726987] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1190.727188] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1190.727379] env[62814]: DEBUG nova.virt.hardware [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1190.733704] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1190.734042] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b19e8c0-7fcb-471a-9ef0-d84c88eeb5b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.753141] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1190.753141] env[62814]: value = "task-4294686" [ 1190.753141] env[62814]: _type = "Task" [ 1190.753141] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.761458] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294686, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.858761] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 9683894b-a300-4400-a1b9-db62478f42c5] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1190.946091] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1190.946316] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1190.947181] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9d61ca-d4af-476d-ae18-7385fc555801 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.967887] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493cd111-28d9-41a3-8def-bc12d78f8097 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.999098] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Reconfiguring VM to detach interface {{(pid=62814) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1190.999098] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d77c58fc-ab79-4f2c-8472-1cbcb7642674 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.018722] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1191.018722] env[62814]: value = "task-4294687" [ 1191.018722] env[62814]: _type = "Task" [ 1191.018722] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.027931] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.028932] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.029330] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1191.029330] env[62814]: DEBUG nova.network.neutron [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1191.129937] env[62814]: DEBUG nova.scheduler.client.report [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1191.210751] env[62814]: DEBUG oslo_vmware.api [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294684, 'name': ReconfigVM_Task, 'duration_secs': 1.347119} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.211050] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Reconfigured VM instance instance-00000068 to attach disk [datastore1] volume-c2969fdc-ffa5-4fde-b88c-26966e10e336/volume-c2969fdc-ffa5-4fde-b88c-26966e10e336.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1191.216033] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae6a18f1-552b-409a-92cf-274ff82c0f4a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.230325] env[62814]: DEBUG oslo_vmware.api [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1191.230325] env[62814]: value = "task-4294688" [ 1191.230325] env[62814]: _type = "Task" [ 1191.230325] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.238518] env[62814]: DEBUG oslo_vmware.api [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294688, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.262387] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294686, 'name': ReconfigVM_Task, 'duration_secs': 0.451975} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.262654] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1191.263425] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5098ca82-8239-4805-aa23-bd781b47385a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.284987] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 40e3c00b-2129-476e-bc67-a0ef13781990/40e3c00b-2129-476e-bc67-a0ef13781990.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1191.285623] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d64702c3-bd58-4536-894f-5fc108219c34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.303383] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1191.303383] env[62814]: value = "task-4294689" [ 1191.303383] env[62814]: _type = "Task" [ 1191.303383] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1191.315335] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294689, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.362040] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: c8765756-0870-4a06-a1a5-d02177959b29] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1191.528695] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.613965] env[62814]: DEBUG nova.compute.manager [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received event network-vif-plugged-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1191.614299] env[62814]: DEBUG oslo_concurrency.lockutils [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] Acquiring lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1191.614601] env[62814]: DEBUG oslo_concurrency.lockutils [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] Lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.614877] env[62814]: DEBUG oslo_concurrency.lockutils [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] Lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.615200] env[62814]: DEBUG nova.compute.manager [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] No waiting events found dispatching network-vif-plugged-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1191.615383] env[62814]: WARNING nova.compute.manager [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received unexpected event network-vif-plugged-a7104ea3-596d-4d10-952c-ac861087e580 for instance with vm_state shelved_offloaded and task_state spawning. [ 1191.615612] env[62814]: DEBUG nova.compute.manager [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received event network-changed-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1191.615837] env[62814]: DEBUG nova.compute.manager [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Refreshing instance network info cache due to event network-changed-a7104ea3-596d-4d10-952c-ac861087e580. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1191.616093] env[62814]: DEBUG oslo_concurrency.lockutils [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] Acquiring lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.634916] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.257s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1191.635385] env[62814]: DEBUG nova.compute.manager [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1191.637894] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.300s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1191.638131] env[62814]: DEBUG nova.objects.instance [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lazy-loading 'resources' on Instance uuid d65ccde2-b31a-4032-8795-1d609fdfcc73 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1191.739876] env[62814]: DEBUG oslo_vmware.api [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294688, 'name': ReconfigVM_Task, 'duration_secs': 0.14291} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.740206] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845858', 'volume_id': 'c2969fdc-ffa5-4fde-b88c-26966e10e336', 'name': 'volume-c2969fdc-ffa5-4fde-b88c-26966e10e336', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '33ab2d9e-e960-40ae-acf9-4fea0d9c830e', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2969fdc-ffa5-4fde-b88c-26966e10e336', 'serial': 'c2969fdc-ffa5-4fde-b88c-26966e10e336'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1191.815278] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294689, 'name': ReconfigVM_Task, 'duration_secs': 0.325251} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.815562] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 40e3c00b-2129-476e-bc67-a0ef13781990/40e3c00b-2129-476e-bc67-a0ef13781990.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1191.815834] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance '40e3c00b-2129-476e-bc67-a0ef13781990' progress to 50 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1191.839640] env[62814]: DEBUG nova.network.neutron [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating instance_info_cache with network_info: [{"id": "a7104ea3-596d-4d10-952c-ac861087e580", "address": "fa:16:3e:4a:60:cd", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7104ea3-59", "ovs_interfaceid": "a7104ea3-596d-4d10-952c-ac861087e580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1191.864963] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 425e8edd-c002-45a1-bb6f-ee3ac8812509] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1192.030533] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.145038] env[62814]: DEBUG nova.compute.utils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1192.150160] env[62814]: DEBUG nova.compute.manager [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1192.150631] env[62814]: DEBUG nova.network.neutron [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1192.211684] env[62814]: DEBUG nova.policy [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e246f49b0b84cd093549b6d6b384e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6e52480dd2c467790622901940cf385', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1192.324152] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8635fda5-972c-458d-9440-3f40043acf19 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.345266] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1192.349606] env[62814]: DEBUG oslo_concurrency.lockutils [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] Acquired lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1192.349797] env[62814]: DEBUG nova.network.neutron [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Refreshing network info cache for port a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1192.351452] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4bddf39-8f54-40fe-96d1-d84eda59f135 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.375151] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 9cd89da4-d190-4bfc-81e1-da15b98d54ba] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1192.377134] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance '40e3c00b-2129-476e-bc67-a0ef13781990' progress to 67 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1192.399672] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='d46cbb10537ee9e265489a35c69b9cbb',container_format='bare',created_at=2025-06-21T05:43:21Z,direct_url=,disk_format='vmdk',id=d3d599ce-93b8-4dd8-a28d-2a1abc5ea448,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1474705082-shelved',owner='8ed3e3bdfb4d4646a8c9eef1582cde85',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2025-06-21T05:43:36Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1192.399913] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1192.400100] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1192.400311] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1192.400462] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1192.400609] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1192.400843] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1192.401026] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1192.401197] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1192.401364] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1192.401569] env[62814]: DEBUG nova.virt.hardware [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1192.402670] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a22785-db9e-4da8-8ea8-5a30b97d86d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.407409] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d32e4a9-8625-4e51-9f21-9f8f66330740 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.417035] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa024f46-99da-4956-9f04-a9f42083150f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.421849] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd0af22-ce24-46b7-a706-11bf6e3e132e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.435653] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:60:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7104ea3-596d-4d10-952c-ac861087e580', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1192.443058] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1192.467098] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1192.467480] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-36573715-8fa3-4d07-a145-e3fd196c8e55 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.481876] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f42c49b-b0f1-4ad5-9202-f8f5e5307ce3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.490295] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e327fbc-da49-49f5-b5d8-9653d3fb33be {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.494827] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1192.494827] env[62814]: value = "task-4294690" [ 1192.494827] env[62814]: _type = "Task" [ 1192.494827] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1192.507748] env[62814]: DEBUG nova.compute.provider_tree [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1192.513883] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294690, 'name': CreateVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.530609] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1192.587177] env[62814]: DEBUG nova.network.neutron [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Successfully created port: 559f235e-f765-465d-a9be-74bb8d7ed03b {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1192.650958] env[62814]: DEBUG nova.compute.manager [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1192.784305] env[62814]: DEBUG nova.objects.instance [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'flavor' on Instance uuid 33ab2d9e-e960-40ae-acf9-4fea0d9c830e {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1192.891679] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 1f40dc62-1a58-4cfb-8785-c37b68747f37] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1192.927678] env[62814]: DEBUG nova.network.neutron [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Port 580abcb7-b7f2-437a-8a84-738e83af0751 binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1193.005575] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294690, 'name': CreateVM_Task, 'duration_secs': 0.287211} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1193.005757] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1193.006418] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.006621] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1193.006956] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1193.009571] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3634389-6895-4fe0-9b98-0c3aad05f4b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.011620] env[62814]: DEBUG nova.scheduler.client.report [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1193.018032] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1193.018032] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5217af4c-b743-bcfa-c0e7-4d0c2786c9e1" [ 1193.018032] env[62814]: _type = "Task" [ 1193.018032] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.028441] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5217af4c-b743-bcfa-c0e7-4d0c2786c9e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.033411] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.078635] env[62814]: DEBUG nova.network.neutron [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updated VIF entry in instance network info cache for port a7104ea3-596d-4d10-952c-ac861087e580. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1193.079091] env[62814]: DEBUG nova.network.neutron [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating instance_info_cache with network_info: [{"id": "a7104ea3-596d-4d10-952c-ac861087e580", "address": "fa:16:3e:4a:60:cd", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.254", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7104ea3-59", "ovs_interfaceid": "a7104ea3-596d-4d10-952c-ac861087e580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.289727] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a1cc27c2-6b3f-4532-9f31-88170b43c4f3 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.278s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.397684] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: cbc5985a-38e7-4e52-9fb0-264b5cec013a] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1193.516108] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.518518] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.693s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1193.521766] env[62814]: INFO nova.compute.claims [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1193.539023] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.543325] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1193.543597] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Processing image d3d599ce-93b8-4dd8-a28d-2a1abc5ea448 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1193.543996] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1193.544102] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1193.544296] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1193.544576] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5aecfd2a-0f35-4b83-a81e-3f1534810eef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.554026] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1193.554194] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1193.554945] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ab81040-c53e-418e-aee1-e289bf732ddf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.558534] env[62814]: INFO nova.scheduler.client.report [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Deleted allocations for instance d65ccde2-b31a-4032-8795-1d609fdfcc73 [ 1193.563865] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1193.563865] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526a5055-1465-114a-bb07-89b2e6a1bf94" [ 1193.563865] env[62814]: _type = "Task" [ 1193.563865] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1193.572075] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526a5055-1465-114a-bb07-89b2e6a1bf94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1193.581549] env[62814]: DEBUG oslo_concurrency.lockutils [req-1327f831-09db-4dc3-9848-6549d6b51ba2 req-4146d0d7-7129-430d-af38-65d9152854e7 service nova] Releasing lock "refresh_cache-1f13051e-e51b-4981-9445-d5420c3c1818" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1193.660783] env[62814]: DEBUG nova.compute.manager [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1193.687491] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1193.687730] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1193.687917] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1193.688127] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1193.688275] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1193.688425] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1193.688634] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1193.688792] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1193.688957] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1193.689131] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1193.689305] env[62814]: DEBUG nova.virt.hardware [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1193.690174] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1db2064-aa16-4a44-994e-4fee6bac828c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.698517] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee649ad7-102f-47da-b718-e72dc9c749ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.901158] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 3e4479a0-8edd-4b37-8cc9-2c91275b88ee] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1193.953489] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "40e3c00b-2129-476e-bc67-a0ef13781990-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1193.953489] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "40e3c00b-2129-476e-bc67-a0ef13781990-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1193.953685] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "40e3c00b-2129-476e-bc67-a0ef13781990-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.991074] env[62814]: DEBUG nova.compute.manager [req-b096213b-7ce6-469f-aabc-3c561443659c req-19195b88-c3a6-4613-9408-988761ebbb46 service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Received event network-vif-plugged-559f235e-f765-465d-a9be-74bb8d7ed03b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1193.991306] env[62814]: DEBUG oslo_concurrency.lockutils [req-b096213b-7ce6-469f-aabc-3c561443659c req-19195b88-c3a6-4613-9408-988761ebbb46 service nova] Acquiring lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1193.991510] env[62814]: DEBUG oslo_concurrency.lockutils [req-b096213b-7ce6-469f-aabc-3c561443659c req-19195b88-c3a6-4613-9408-988761ebbb46 service nova] Lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1193.991677] env[62814]: DEBUG oslo_concurrency.lockutils [req-b096213b-7ce6-469f-aabc-3c561443659c req-19195b88-c3a6-4613-9408-988761ebbb46 service nova] Lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1193.991882] env[62814]: DEBUG nova.compute.manager [req-b096213b-7ce6-469f-aabc-3c561443659c req-19195b88-c3a6-4613-9408-988761ebbb46 service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] No waiting events found dispatching network-vif-plugged-559f235e-f765-465d-a9be-74bb8d7ed03b {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1193.992077] env[62814]: WARNING nova.compute.manager [req-b096213b-7ce6-469f-aabc-3c561443659c req-19195b88-c3a6-4613-9408-988761ebbb46 service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Received unexpected event network-vif-plugged-559f235e-f765-465d-a9be-74bb8d7ed03b for instance with vm_state building and task_state spawning. [ 1194.033925] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.069255] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5b9a7506-2ef9-4b54-a38c-51b10879ef4f tempest-ServerShowV247Test-1879384097 tempest-ServerShowV247Test-1879384097-project-member] Lock "d65ccde2-b31a-4032-8795-1d609fdfcc73" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.585s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1194.077150] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Preparing fetch location {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1194.077412] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Fetch image to [datastore2] OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf/OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf.vmdk {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1194.077559] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Downloading stream optimized image d3d599ce-93b8-4dd8-a28d-2a1abc5ea448 to [datastore2] OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf/OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf.vmdk on the data store datastore2 as vApp {{(pid=62814) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1194.077727] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Downloading image file data d3d599ce-93b8-4dd8-a28d-2a1abc5ea448 to the ESX as VM named 'OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf' {{(pid=62814) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1194.080960] env[62814]: DEBUG nova.network.neutron [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Successfully updated port: 559f235e-f765-465d-a9be-74bb8d7ed03b {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1194.149123] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1194.149123] env[62814]: value = "resgroup-9" [ 1194.149123] env[62814]: _type = "ResourcePool" [ 1194.149123] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1194.149406] env[62814]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-091bfd68-6572-4b74-8faa-8cfdab5c55d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.171100] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lease: (returnval){ [ 1194.171100] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522d86d0-8fe0-b298-2ac6-1fc918a639ce" [ 1194.171100] env[62814]: _type = "HttpNfcLease" [ 1194.171100] env[62814]: } obtained for vApp import into resource pool (val){ [ 1194.171100] env[62814]: value = "resgroup-9" [ 1194.171100] env[62814]: _type = "ResourcePool" [ 1194.171100] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1194.171400] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the lease: (returnval){ [ 1194.171400] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522d86d0-8fe0-b298-2ac6-1fc918a639ce" [ 1194.171400] env[62814]: _type = "HttpNfcLease" [ 1194.171400] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1194.177428] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1194.177428] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522d86d0-8fe0-b298-2ac6-1fc918a639ce" [ 1194.177428] env[62814]: _type = "HttpNfcLease" [ 1194.177428] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1194.404306] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: d366a755-49b4-427b-8564-d8572a7fbbb7] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1194.472329] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1194.472617] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1194.533100] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1194.584480] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "refresh_cache-44d94dc4-d7b3-4799-9044-9c1ecc80c88c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1194.584480] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "refresh_cache-44d94dc4-d7b3-4799-9044-9c1ecc80c88c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1194.584636] env[62814]: DEBUG nova.network.neutron [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1194.681285] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1194.681285] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522d86d0-8fe0-b298-2ac6-1fc918a639ce" [ 1194.681285] env[62814]: _type = "HttpNfcLease" [ 1194.681285] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1194.681839] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1194.681839] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]522d86d0-8fe0-b298-2ac6-1fc918a639ce" [ 1194.681839] env[62814]: _type = "HttpNfcLease" [ 1194.681839] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1194.682318] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79944cfe-29cb-47ed-a565-2ebb86763110 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.689345] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524201ed-c60c-82c1-43a7-924922f68670/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1194.689523] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524201ed-c60c-82c1-43a7-924922f68670/disk-0.vmdk. {{(pid=62814) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1194.751278] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1fb4eabf-514b-4af4-bed6-836a7895f594 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.774925] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b9b28b-5889-4837-a33e-54b3347a4438 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.782741] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047f0f67-9f65-4277-924f-31158e1f2778 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.816402] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd56983c-45c4-4ca2-a34d-d3533e8cf2da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.824144] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44505095-520d-4858-abb1-71071241dae0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.838017] env[62814]: DEBUG nova.compute.provider_tree [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1194.907560] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 9343301c-dfe9-41b0-b4a0-067af544d297] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1194.976794] env[62814]: DEBUG nova.compute.utils [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1195.029280] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1195.029497] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1195.029722] env[62814]: DEBUG nova.network.neutron [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1195.038435] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.119396] env[62814]: DEBUG nova.network.neutron [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1195.123831] env[62814]: DEBUG oslo_concurrency.lockutils [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.124169] env[62814]: DEBUG oslo_concurrency.lockutils [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.124437] env[62814]: DEBUG oslo_concurrency.lockutils [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1195.124714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.124879] env[62814]: DEBUG oslo_concurrency.lockutils [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.126881] env[62814]: INFO nova.compute.manager [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Terminating instance [ 1195.292316] env[62814]: DEBUG nova.network.neutron [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Updating instance_info_cache with network_info: [{"id": "559f235e-f765-465d-a9be-74bb8d7ed03b", "address": "fa:16:3e:16:4b:c1", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559f235e-f7", "ovs_interfaceid": "559f235e-f765-465d-a9be-74bb8d7ed03b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.341069] env[62814]: DEBUG nova.scheduler.client.report [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1195.410900] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: bd0933ca-aab7-4dd4-a570-1a58a720f377] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1195.480832] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.538475] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.632351] env[62814]: DEBUG nova.compute.manager [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1195.632659] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1195.633770] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe7e9f5-33a0-4e12-a2ec-f6bfeaad1821 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.645261] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1195.645518] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a34f0f4-0023-4986-b39b-b4849395415c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.656725] env[62814]: DEBUG oslo_vmware.api [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1195.656725] env[62814]: value = "task-4294692" [ 1195.656725] env[62814]: _type = "Task" [ 1195.656725] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.667761] env[62814]: DEBUG oslo_vmware.api [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294692, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.794935] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "refresh_cache-44d94dc4-d7b3-4799-9044-9c1ecc80c88c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1195.795423] env[62814]: DEBUG nova.compute.manager [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Instance network_info: |[{"id": "559f235e-f765-465d-a9be-74bb8d7ed03b", "address": "fa:16:3e:16:4b:c1", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559f235e-f7", "ovs_interfaceid": "559f235e-f765-465d-a9be-74bb8d7ed03b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1195.795875] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:4b:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '51bac3c3-00ab-4a07-9e28-b3c951dee565', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '559f235e-f765-465d-a9be-74bb8d7ed03b', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1195.803982] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1195.804317] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1195.804380] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cf8b4dc7-4712-4b7b-8ea4-017caafe3107 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.831193] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1195.831193] env[62814]: value = "task-4294693" [ 1195.831193] env[62814]: _type = "Task" [ 1195.831193] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1195.840971] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294693, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.846935] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.328s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1195.847513] env[62814]: DEBUG nova.compute.manager [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1195.850445] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.975s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1195.864996] env[62814]: DEBUG nova.network.neutron [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance_info_cache with network_info: [{"id": "580abcb7-b7f2-437a-8a84-738e83af0751", "address": "fa:16:3e:f1:6c:e8", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap580abcb7-b7", "ovs_interfaceid": "580abcb7-b7f2-437a-8a84-738e83af0751", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1195.915718] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: f5ad9d70-75fb-4881-8853-5ede4d0903f2] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1195.921404] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Completed reading data from the image iterator. {{(pid=62814) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1195.921861] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524201ed-c60c-82c1-43a7-924922f68670/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1195.922730] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42aded7-4391-46e7-ad47-4b816a0207d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.931123] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524201ed-c60c-82c1-43a7-924922f68670/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1195.931334] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524201ed-c60c-82c1-43a7-924922f68670/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1195.931605] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-35ee68e7-3902-4240-a89f-cd5ef57ebaec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.034896] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.073983] env[62814]: DEBUG nova.compute.manager [req-a84684da-d54c-4e92-9395-8ef280713f77 req-e15c195a-a8b4-4b1d-ba60-7f93f5db3ada service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Received event network-changed-559f235e-f765-465d-a9be-74bb8d7ed03b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1196.074259] env[62814]: DEBUG nova.compute.manager [req-a84684da-d54c-4e92-9395-8ef280713f77 req-e15c195a-a8b4-4b1d-ba60-7f93f5db3ada service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Refreshing instance network info cache due to event network-changed-559f235e-f765-465d-a9be-74bb8d7ed03b. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1196.074523] env[62814]: DEBUG oslo_concurrency.lockutils [req-a84684da-d54c-4e92-9395-8ef280713f77 req-e15c195a-a8b4-4b1d-ba60-7f93f5db3ada service nova] Acquiring lock "refresh_cache-44d94dc4-d7b3-4799-9044-9c1ecc80c88c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.074763] env[62814]: DEBUG oslo_concurrency.lockutils [req-a84684da-d54c-4e92-9395-8ef280713f77 req-e15c195a-a8b4-4b1d-ba60-7f93f5db3ada service nova] Acquired lock "refresh_cache-44d94dc4-d7b3-4799-9044-9c1ecc80c88c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1196.074933] env[62814]: DEBUG nova.network.neutron [req-a84684da-d54c-4e92-9395-8ef280713f77 req-e15c195a-a8b4-4b1d-ba60-7f93f5db3ada service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Refreshing network info cache for port 559f235e-f765-465d-a9be-74bb8d7ed03b {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1196.169134] env[62814]: DEBUG oslo_vmware.api [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294692, 'name': PowerOffVM_Task, 'duration_secs': 0.173904} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.170465] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1196.170707] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1196.171148] env[62814]: DEBUG oslo_vmware.rw_handles [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524201ed-c60c-82c1-43a7-924922f68670/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1196.171338] env[62814]: INFO nova.virt.vmwareapi.images [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Downloaded image file data d3d599ce-93b8-4dd8-a28d-2a1abc5ea448 [ 1196.171607] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fc04f63b-35ce-4903-b70e-c5fee2c6c5c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.173597] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b800c38e-736a-4af1-aed0-2c5c36f3f579 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.189997] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-170c7a9b-43d5-4cb7-bc95-5bb2a256b65e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.223164] env[62814]: INFO nova.virt.vmwareapi.images [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] The imported VM was unregistered [ 1196.225787] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Caching image {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1196.226034] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating directory with path [datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448 {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1196.226313] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e87c3f05-3a61-45e5-b860-bd4572c37d43 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.235247] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1196.235451] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1196.235631] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleting the datastore file [datastore2] edcdb1e5-ed9d-49e6-97e6-ea7629682547 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1196.235874] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c23a090c-7dba-48e3-8786-aea5f58c7334 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.238503] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Created directory with path [datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448 {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1196.238682] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf/OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf.vmdk to [datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448.vmdk. {{(pid=62814) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1196.239224] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-5b14dabc-e4e5-41b1-aa34-5096937112fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.243803] env[62814]: DEBUG oslo_vmware.api [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for the task: (returnval){ [ 1196.243803] env[62814]: value = "task-4294696" [ 1196.243803] env[62814]: _type = "Task" [ 1196.243803] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.247528] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1196.247528] env[62814]: value = "task-4294697" [ 1196.247528] env[62814]: _type = "Task" [ 1196.247528] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.253486] env[62814]: DEBUG oslo_vmware.api [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294696, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.258076] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294697, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.341529] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294693, 'name': CreateVM_Task, 'duration_secs': 0.344703} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.341714] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1196.342364] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.342538] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1196.342858] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1196.343128] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c862e07-2d6a-4b80-a080-bc478e4cce82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.348313] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1196.348313] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52de14e1-972d-ef34-f5ab-4c3c96c5f32c" [ 1196.348313] env[62814]: _type = "Task" [ 1196.348313] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.354242] env[62814]: DEBUG nova.compute.utils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1196.357899] env[62814]: INFO nova.compute.claims [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1196.364716] env[62814]: DEBUG nova.compute.manager [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1196.364906] env[62814]: DEBUG nova.network.neutron [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1196.366770] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52de14e1-972d-ef34-f5ab-4c3c96c5f32c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.370028] env[62814]: DEBUG oslo_concurrency.lockutils [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1196.404731] env[62814]: DEBUG nova.policy [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd05b1b390e0489d91f4f386f4c6fe41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '65a886f76f404ce5a6d0841241d8f120', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1196.422556] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 4ed66b36-b6c6-4673-9c03-169a01134574] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1196.539860] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.555354] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1196.555738] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1196.556074] env[62814]: INFO nova.compute.manager [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Attaching volume ea4badfa-3567-42a6-83c7-13ecccee11d3 to /dev/sdc [ 1196.601018] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8cb5468-edc0-481b-b195-2b8432555770 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.606975] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646e9317-53f6-4c0b-b629-ebdd2c3a9ff7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.622285] env[62814]: DEBUG nova.virt.block_device [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Updating existing volume attachment record: 9acce338-7857-4b98-bab1-f05980852a67 {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1196.762214] env[62814]: DEBUG oslo_vmware.api [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Task: {'id': task-4294696, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.275193} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.765589] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1196.765861] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1196.765972] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1196.766209] env[62814]: INFO nova.compute.manager [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1196.766466] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1196.766674] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294697, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.766889] env[62814]: DEBUG nova.compute.manager [-] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1196.766989] env[62814]: DEBUG nova.network.neutron [-] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1196.861633] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52de14e1-972d-ef34-f5ab-4c3c96c5f32c, 'name': SearchDatastore_Task, 'duration_secs': 0.032738} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1196.862224] env[62814]: DEBUG nova.compute.manager [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1196.868034] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1196.868034] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1196.868034] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.868034] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1196.868034] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1196.872022] env[62814]: INFO nova.compute.resource_tracker [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating resource usage from migration 7c79fbaa-6c2c-4243-8ae4-fa861c910052 [ 1196.876048] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2abfa7f-810e-4d8b-8316-23148cc53dfe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.900809] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1196.900972] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1196.902034] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb42f85b-7115-48f5-b7fc-b0c71806e999 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.909311] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1196.909311] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52397e08-cb50-546d-109d-5b58336c3011" [ 1196.909311] env[62814]: _type = "Task" [ 1196.909311] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1196.923376] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a2cca57-08f2-4657-aa87-a7e54ebf150f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.930380] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 59d21ef7-df97-49ac-9329-4c18df6dd087] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1196.932456] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52397e08-cb50-546d-109d-5b58336c3011, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1196.962512] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb816ac2-c6e7-435a-976e-d31477ec076a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1196.969931] env[62814]: DEBUG nova.network.neutron [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Successfully created port: e07ce73b-c8ba-4a79-a9a4-a5d1672041d9 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1196.975592] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance '40e3c00b-2129-476e-bc67-a0ef13781990' progress to 83 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1196.980474] env[62814]: DEBUG nova.network.neutron [req-a84684da-d54c-4e92-9395-8ef280713f77 req-e15c195a-a8b4-4b1d-ba60-7f93f5db3ada service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Updated VIF entry in instance network info cache for port 559f235e-f765-465d-a9be-74bb8d7ed03b. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1196.980847] env[62814]: DEBUG nova.network.neutron [req-a84684da-d54c-4e92-9395-8ef280713f77 req-e15c195a-a8b4-4b1d-ba60-7f93f5db3ada service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Updating instance_info_cache with network_info: [{"id": "559f235e-f765-465d-a9be-74bb8d7ed03b", "address": "fa:16:3e:16:4b:c1", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap559f235e-f7", "ovs_interfaceid": "559f235e-f765-465d-a9be-74bb8d7ed03b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1197.040914] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.140801] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16787e87-3aa3-4c41-b607-ce515c1a1ccf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.151648] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe8d1c5d-c879-4974-83d0-e958be564237 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.189257] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d4d151-65d4-4e5f-915e-9e506afb25c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.199429] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8622bed-6ea9-4457-a959-a1655f0e1468 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.218611] env[62814]: DEBUG nova.compute.provider_tree [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1197.263680] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294697, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.423258] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52397e08-cb50-546d-109d-5b58336c3011, 'name': SearchDatastore_Task, 'duration_secs': 0.092412} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.424226] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5499c132-55d2-4d7f-ac55-f7e68f38ec74 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.432414] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1197.432414] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528e412b-c48e-ad08-9d8f-80db44ff2caa" [ 1197.432414] env[62814]: _type = "Task" [ 1197.432414] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.437352] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: d9242042-6209-4b04-bf00-00dd04d9d6a0] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1197.446774] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528e412b-c48e-ad08-9d8f-80db44ff2caa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.486140] env[62814]: DEBUG oslo_concurrency.lockutils [req-a84684da-d54c-4e92-9395-8ef280713f77 req-e15c195a-a8b4-4b1d-ba60-7f93f5db3ada service nova] Releasing lock "refresh_cache-44d94dc4-d7b3-4799-9044-9c1ecc80c88c" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1197.488038] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1197.488332] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-977878f8-ec41-49b8-8fe2-33c9d1e4d442 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.497437] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1197.497437] env[62814]: value = "task-4294699" [ 1197.497437] env[62814]: _type = "Task" [ 1197.497437] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.510716] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294699, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.544459] env[62814]: DEBUG oslo_vmware.api [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294687, 'name': ReconfigVM_Task, 'duration_secs': 6.098367} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.545847] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1197.546171] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Reconfigured VM to detach interface {{(pid=62814) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1197.549753] env[62814]: DEBUG nova.compute.manager [req-ec2b26c3-2e3e-4f68-8a21-81549fc13702 req-1cad02be-b16d-4332-ae8f-39e11e13d262 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Received event network-vif-deleted-9950de47-d55a-480b-b88e-818e09b9bc80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1197.549869] env[62814]: INFO nova.compute.manager [req-ec2b26c3-2e3e-4f68-8a21-81549fc13702 req-1cad02be-b16d-4332-ae8f-39e11e13d262 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Neutron deleted interface 9950de47-d55a-480b-b88e-818e09b9bc80; detaching it from the instance and deleting it from the info cache [ 1197.550061] env[62814]: DEBUG nova.network.neutron [req-ec2b26c3-2e3e-4f68-8a21-81549fc13702 req-1cad02be-b16d-4332-ae8f-39e11e13d262 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1197.743611] env[62814]: ERROR nova.scheduler.client.report [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [req-77b8df4e-be16-4e11-9535-ec8e8835d12d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-77b8df4e-be16-4e11-9535-ec8e8835d12d"}]} [ 1197.765574] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294697, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1197.768910] env[62814]: DEBUG nova.scheduler.client.report [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1197.783905] env[62814]: DEBUG nova.scheduler.client.report [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1197.784173] env[62814]: DEBUG nova.compute.provider_tree [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1197.802231] env[62814]: DEBUG nova.scheduler.client.report [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1197.823095] env[62814]: DEBUG nova.scheduler.client.report [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1197.875149] env[62814]: DEBUG nova.compute.manager [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1197.903498] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1197.904086] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1197.904086] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1197.904222] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1197.904272] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1197.904487] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1197.904949] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1197.904949] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1197.905052] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1197.905193] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1197.905370] env[62814]: DEBUG nova.virt.hardware [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1197.906619] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2da2dde5-74ad-48f7-b942-416e6bccfb7d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.917862] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae0f0ea-4b22-4b3b-ba51-2704563fce66 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.945818] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: c7cef7f3-11db-44e1-a454-98830b465b52] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1197.956557] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528e412b-c48e-ad08-9d8f-80db44ff2caa, 'name': SearchDatastore_Task, 'duration_secs': 0.090814} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1197.956851] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1197.957153] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 44d94dc4-d7b3-4799-9044-9c1ecc80c88c/44d94dc4-d7b3-4799-9044-9c1ecc80c88c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1197.957439] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4682d999-cc82-4c01-aaf1-92279282ebed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1197.967166] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1197.967166] env[62814]: value = "task-4294700" [ 1197.967166] env[62814]: _type = "Task" [ 1197.967166] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1197.975878] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294700, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.009521] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294699, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.018930] env[62814]: DEBUG nova.network.neutron [-] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.043967] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f61042-658b-4f20-8910-1d25d2e5270c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.052358] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f5296b-8e81-4292-9664-5706db5e4466 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.056099] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c6f31603-27fa-4961-a399-749df7be88b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.090131] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06ca859-65dd-47cb-929e-3ad16db1d55c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.095541] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943cf011-31a1-4265-b8f2-ce72ccaf91d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.114444] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e0a2c1-77af-4896-bb1b-9156df72e3b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.128277] env[62814]: DEBUG nova.compute.manager [req-ec2b26c3-2e3e-4f68-8a21-81549fc13702 req-1cad02be-b16d-4332-ae8f-39e11e13d262 service nova] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Detach interface failed, port_id=9950de47-d55a-480b-b88e-818e09b9bc80, reason: Instance edcdb1e5-ed9d-49e6-97e6-ea7629682547 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1198.140722] env[62814]: DEBUG nova.compute.provider_tree [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1198.265331] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294697, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.452715] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 6a592192-1b41-4be2-84a6-c3b76a4e5643] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1198.481730] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294700, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.510652] env[62814]: DEBUG oslo_vmware.api [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294699, 'name': PowerOnVM_Task, 'duration_secs': 0.802644} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.510964] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1198.511167] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-621da3f6-726b-4eff-a3bb-7a002131670b tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance '40e3c00b-2129-476e-bc67-a0ef13781990' progress to 100 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1198.521612] env[62814]: INFO nova.compute.manager [-] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Took 1.75 seconds to deallocate network for instance. [ 1198.598454] env[62814]: DEBUG nova.compute.manager [req-3c5f8e35-3eb0-4e1d-a526-99eae070142a req-c7848572-ade4-4b03-92de-bde9a8af3ad9 service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Received event network-vif-plugged-e07ce73b-c8ba-4a79-a9a4-a5d1672041d9 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1198.598677] env[62814]: DEBUG oslo_concurrency.lockutils [req-3c5f8e35-3eb0-4e1d-a526-99eae070142a req-c7848572-ade4-4b03-92de-bde9a8af3ad9 service nova] Acquiring lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1198.598885] env[62814]: DEBUG oslo_concurrency.lockutils [req-3c5f8e35-3eb0-4e1d-a526-99eae070142a req-c7848572-ade4-4b03-92de-bde9a8af3ad9 service nova] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1198.599074] env[62814]: DEBUG oslo_concurrency.lockutils [req-3c5f8e35-3eb0-4e1d-a526-99eae070142a req-c7848572-ade4-4b03-92de-bde9a8af3ad9 service nova] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1198.599309] env[62814]: DEBUG nova.compute.manager [req-3c5f8e35-3eb0-4e1d-a526-99eae070142a req-c7848572-ade4-4b03-92de-bde9a8af3ad9 service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] No waiting events found dispatching network-vif-plugged-e07ce73b-c8ba-4a79-a9a4-a5d1672041d9 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1198.599489] env[62814]: WARNING nova.compute.manager [req-3c5f8e35-3eb0-4e1d-a526-99eae070142a req-c7848572-ade4-4b03-92de-bde9a8af3ad9 service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Received unexpected event network-vif-plugged-e07ce73b-c8ba-4a79-a9a4-a5d1672041d9 for instance with vm_state building and task_state spawning. [ 1198.680578] env[62814]: DEBUG nova.scheduler.client.report [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 171 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1198.680845] env[62814]: DEBUG nova.compute.provider_tree [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 171 to 172 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1198.681036] env[62814]: DEBUG nova.compute.provider_tree [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1198.734275] env[62814]: DEBUG nova.network.neutron [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Successfully updated port: e07ce73b-c8ba-4a79-a9a4-a5d1672041d9 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1198.766170] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294697, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.509511} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1198.766567] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf/OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf.vmdk to [datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448.vmdk. [ 1198.766839] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Cleaning up location [datastore2] OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1198.767100] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_a26c1564-314b-45f4-bc2b-3b03d8c917bf {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1198.767422] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6ede948a-392a-4019-9560-fd5724d6c4c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.774508] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1198.774508] env[62814]: value = "task-4294701" [ 1198.774508] env[62814]: _type = "Task" [ 1198.774508] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1198.785356] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294701, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1198.886944] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1198.886944] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1198.886944] env[62814]: DEBUG nova.network.neutron [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1198.958245] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1198.958245] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Cleaning up deleted instances with incomplete migration {{(pid=62814) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11903}} [ 1198.985506] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294700, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.029653] env[62814]: DEBUG oslo_concurrency.lockutils [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.187486] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.337s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1199.187875] env[62814]: INFO nova.compute.manager [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Migrating [ 1199.196360] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 9.219s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.237217] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "refresh_cache-0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.237385] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "refresh_cache-0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1199.238053] env[62814]: DEBUG nova.network.neutron [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1199.284625] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294701, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162609} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.284839] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1199.284935] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1199.285278] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448.vmdk to [datastore2] 1f13051e-e51b-4981-9445-d5420c3c1818/1f13051e-e51b-4981-9445-d5420c3c1818.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1199.287026] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-504a0276-8025-4dee-9364-a2595a9accb2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.291438] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1199.291438] env[62814]: value = "task-4294703" [ 1199.291438] env[62814]: _type = "Task" [ 1199.291438] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.298998] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.460810] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1199.483607] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294700, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.144646} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1199.483923] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 44d94dc4-d7b3-4799-9044-9c1ecc80c88c/44d94dc4-d7b3-4799-9044-9c1ecc80c88c.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1199.484232] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1199.484532] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43fb6a59-32b5-455d-8bd5-0804e9bac313 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.491206] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1199.491206] env[62814]: value = "task-4294704" [ 1199.491206] env[62814]: _type = "Task" [ 1199.491206] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.504447] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1199.585789] env[62814]: DEBUG nova.compute.manager [req-d71bca2c-6950-41a0-a3ac-bcbc41237ac8 req-943c67db-2894-4fd8-8770-f755c20939bc service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received event network-vif-deleted-88931aee-29b6-4f58-9ee6-0210612c8211 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1199.585985] env[62814]: INFO nova.compute.manager [req-d71bca2c-6950-41a0-a3ac-bcbc41237ac8 req-943c67db-2894-4fd8-8770-f755c20939bc service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Neutron deleted interface 88931aee-29b6-4f58-9ee6-0210612c8211; detaching it from the instance and deleting it from the info cache [ 1199.586299] env[62814]: DEBUG nova.network.neutron [req-d71bca2c-6950-41a0-a3ac-bcbc41237ac8 req-943c67db-2894-4fd8-8770-f755c20939bc service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [{"id": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "address": "fa:16:3e:e8:10:0d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c6b72d7-d1", "ovs_interfaceid": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "fec67180-c78f-4b6c-b106-a0d1ab127ec4", "address": "fa:16:3e:00:36:d2", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfec67180-c7", "ovs_interfaceid": "fec67180-c78f-4b6c-b106-a0d1ab127ec4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.665488] env[62814]: INFO nova.network.neutron [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Port 88931aee-29b6-4f58-9ee6-0210612c8211 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1199.665699] env[62814]: INFO nova.network.neutron [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Port fec67180-c78f-4b6c-b106-a0d1ab127ec4 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1199.666124] env[62814]: DEBUG nova.network.neutron [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [{"id": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "address": "fa:16:3e:e8:10:0d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.181", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1c6b72d7-d1", "ovs_interfaceid": "1c6b72d7-d148-4280-981c-9bfbf768bd3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1199.708912] env[62814]: INFO nova.compute.claims [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1199.712608] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.712844] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.713062] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1199.713251] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1199.713418] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1199.715182] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1199.715367] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1199.715540] env[62814]: DEBUG nova.network.neutron [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1199.717855] env[62814]: INFO nova.compute.manager [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Terminating instance [ 1199.797091] env[62814]: DEBUG nova.network.neutron [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1199.806769] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294703, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.003885] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108934} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.004254] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1200.005539] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649f9096-36e5-4a17-81be-4f8feead7bee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.008737] env[62814]: DEBUG nova.network.neutron [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Updating instance_info_cache with network_info: [{"id": "e07ce73b-c8ba-4a79-a9a4-a5d1672041d9", "address": "fa:16:3e:86:21:ad", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape07ce73b-c8", "ovs_interfaceid": "e07ce73b-c8ba-4a79-a9a4-a5d1672041d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.036009] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 44d94dc4-d7b3-4799-9044-9c1ecc80c88c/44d94dc4-d7b3-4799-9044-9c1ecc80c88c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1200.044353] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6694176-62a5-4be4-8490-73d4e45d21d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.069326] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1200.069326] env[62814]: value = "task-4294705" [ 1200.069326] env[62814]: _type = "Task" [ 1200.069326] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.080212] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294705, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.089852] env[62814]: DEBUG oslo_concurrency.lockutils [req-d71bca2c-6950-41a0-a3ac-bcbc41237ac8 req-943c67db-2894-4fd8-8770-f755c20939bc service nova] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.170116] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1200.218570] env[62814]: INFO nova.compute.resource_tracker [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating resource usage from migration 7a60a3b4-92a0-49df-b1ed-bf600dddbef6 [ 1200.224058] env[62814]: DEBUG nova.compute.manager [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1200.224324] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1200.225579] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17fba0ef-9de3-4b19-8481-62af178ec062 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.233908] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1200.234172] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f44ba1d-5123-49ae-9099-3b8f8b2b1562 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.240850] env[62814]: DEBUG oslo_vmware.api [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1200.240850] env[62814]: value = "task-4294706" [ 1200.240850] env[62814]: _type = "Task" [ 1200.240850] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.253116] env[62814]: DEBUG oslo_vmware.api [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294706, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.304891] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294703, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.436472] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "40e3c00b-2129-476e-bc67-a0ef13781990" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1200.436731] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "40e3c00b-2129-476e-bc67-a0ef13781990" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1200.436926] env[62814]: DEBUG nova.compute.manager [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Going to confirm migration 6 {{(pid=62814) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 1200.448171] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ead16c59-b6e2-431c-b270-286162d6d8b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.457814] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad21efe5-e44c-4316-bebd-a377679ba78f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.498762] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e55ef9e-5f1e-4400-9bf9-7607780c4c14 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.507727] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44b9dce8-86d9-4097-b9a2-035188105f3d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.513285] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "refresh_cache-0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1200.513602] env[62814]: DEBUG nova.compute.manager [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Instance network_info: |[{"id": "e07ce73b-c8ba-4a79-a9a4-a5d1672041d9", "address": "fa:16:3e:86:21:ad", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape07ce73b-c8", "ovs_interfaceid": "e07ce73b-c8ba-4a79-a9a4-a5d1672041d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1200.514081] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:21:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7e0240aa-a694-48fc-a0f9-6f2d3e71aa12', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e07ce73b-c8ba-4a79-a9a4-a5d1672041d9', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1200.522219] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1200.522987] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1200.523268] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-865a89c8-d44c-45ed-91e0-2e1d6f12a068 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.541255] env[62814]: DEBUG nova.network.neutron [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance_info_cache with network_info: [{"id": "0235e89b-ec0d-4439-817c-dd76d6987b80", "address": "fa:16:3e:46:9d:6d", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0235e89b-ec", "ovs_interfaceid": "0235e89b-ec0d-4439-817c-dd76d6987b80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1200.555047] env[62814]: DEBUG nova.compute.provider_tree [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1200.559473] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1200.559473] env[62814]: value = "task-4294707" [ 1200.559473] env[62814]: _type = "Task" [ 1200.559473] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1200.569518] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294707, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.578909] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.639368] env[62814]: DEBUG nova.compute.manager [req-3ae92f24-a00c-48eb-aa3f-052e538c7629 req-426bbe89-b823-4412-a564-19788993fb9d service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Received event network-changed-e07ce73b-c8ba-4a79-a9a4-a5d1672041d9 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1200.639658] env[62814]: DEBUG nova.compute.manager [req-3ae92f24-a00c-48eb-aa3f-052e538c7629 req-426bbe89-b823-4412-a564-19788993fb9d service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Refreshing instance network info cache due to event network-changed-e07ce73b-c8ba-4a79-a9a4-a5d1672041d9. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1200.639943] env[62814]: DEBUG oslo_concurrency.lockutils [req-3ae92f24-a00c-48eb-aa3f-052e538c7629 req-426bbe89-b823-4412-a564-19788993fb9d service nova] Acquiring lock "refresh_cache-0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.640089] env[62814]: DEBUG oslo_concurrency.lockutils [req-3ae92f24-a00c-48eb-aa3f-052e538c7629 req-426bbe89-b823-4412-a564-19788993fb9d service nova] Acquired lock "refresh_cache-0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1200.640487] env[62814]: DEBUG nova.network.neutron [req-3ae92f24-a00c-48eb-aa3f-052e538c7629 req-426bbe89-b823-4412-a564-19788993fb9d service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Refreshing network info cache for port e07ce73b-c8ba-4a79-a9a4-a5d1672041d9 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1200.674171] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d229c0f6-6f6f-4ba5-a3ee-0cb2075bba02 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-2d7a18f3-456b-470c-a759-632e5d79a8f0-88931aee-29b6-4f58-9ee6-0210612c8211" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.231s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1200.753900] env[62814]: DEBUG oslo_vmware.api [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294706, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.803133] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294703, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.958738] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.958738] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1200.980987] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1200.981225] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1200.981432] env[62814]: DEBUG nova.network.neutron [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1200.981625] env[62814]: DEBUG nova.objects.instance [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lazy-loading 'info_cache' on Instance uuid 40e3c00b-2129-476e-bc67-a0ef13781990 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.044061] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1201.071649] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294707, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.080573] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.090761] env[62814]: DEBUG nova.scheduler.client.report [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 172 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1201.091031] env[62814]: DEBUG nova.compute.provider_tree [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 172 to 173 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1201.091222] env[62814]: DEBUG nova.compute.provider_tree [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1201.185966] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1201.186341] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845862', 'volume_id': 'ea4badfa-3567-42a6-83c7-13ecccee11d3', 'name': 'volume-ea4badfa-3567-42a6-83c7-13ecccee11d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '33ab2d9e-e960-40ae-acf9-4fea0d9c830e', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea4badfa-3567-42a6-83c7-13ecccee11d3', 'serial': 'ea4badfa-3567-42a6-83c7-13ecccee11d3'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1201.187409] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8033f44e-c1f3-42f3-b070-a10310639326 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.212576] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03737ee0-9e13-43db-baac-584de8641b02 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.246375] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] volume-ea4badfa-3567-42a6-83c7-13ecccee11d3/volume-ea4badfa-3567-42a6-83c7-13ecccee11d3.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1201.249795] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-877dcba4-284d-438a-8706-4b454d90ab62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.275488] env[62814]: DEBUG oslo_vmware.api [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294706, 'name': PowerOffVM_Task, 'duration_secs': 1.002338} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.277024] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1201.277263] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1201.277706] env[62814]: DEBUG oslo_vmware.api [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1201.277706] env[62814]: value = "task-4294708" [ 1201.277706] env[62814]: _type = "Task" [ 1201.277706] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.277879] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dae9399e-f06c-425c-b117-d4be6b408e46 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.289276] env[62814]: DEBUG oslo_vmware.api [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294708, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.307246] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294703, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.429369] env[62814]: DEBUG nova.network.neutron [req-3ae92f24-a00c-48eb-aa3f-052e538c7629 req-426bbe89-b823-4412-a564-19788993fb9d service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Updated VIF entry in instance network info cache for port e07ce73b-c8ba-4a79-a9a4-a5d1672041d9. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1201.429791] env[62814]: DEBUG nova.network.neutron [req-3ae92f24-a00c-48eb-aa3f-052e538c7629 req-426bbe89-b823-4412-a564-19788993fb9d service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Updating instance_info_cache with network_info: [{"id": "e07ce73b-c8ba-4a79-a9a4-a5d1672041d9", "address": "fa:16:3e:86:21:ad", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape07ce73b-c8", "ovs_interfaceid": "e07ce73b-c8ba-4a79-a9a4-a5d1672041d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1201.466436] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.466436] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.466436] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.466436] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.466436] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_power_states {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1201.559273] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1201.559608] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1201.559856] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleting the datastore file [datastore2] 2d7a18f3-456b-470c-a759-632e5d79a8f0 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1201.560218] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-310a2917-9131-4c96-be61-ed26b7c201ce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.573045] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294707, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.578433] env[62814]: DEBUG oslo_vmware.api [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1201.578433] env[62814]: value = "task-4294710" [ 1201.578433] env[62814]: _type = "Task" [ 1201.578433] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.585403] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294705, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.591266] env[62814]: DEBUG oslo_vmware.api [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294710, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.596020] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.400s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1201.596878] env[62814]: INFO nova.compute.manager [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Migrating [ 1201.603197] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.403s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1201.603456] env[62814]: DEBUG nova.objects.instance [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lazy-loading 'resources' on Instance uuid bf2ba498-7c47-4985-a84a-475037deec55 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1201.789840] env[62814]: DEBUG oslo_vmware.api [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294708, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.804887] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294703, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.454686} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1201.805183] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448/d3d599ce-93b8-4dd8-a28d-2a1abc5ea448.vmdk to [datastore2] 1f13051e-e51b-4981-9445-d5420c3c1818/1f13051e-e51b-4981-9445-d5420c3c1818.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1201.805989] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9cf6916-cf2b-4b1d-b371-4cf7e09bf75d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.827922] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 1f13051e-e51b-4981-9445-d5420c3c1818/1f13051e-e51b-4981-9445-d5420c3c1818.vmdk or device None with type streamOptimized {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1201.828218] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-068a246d-9a59-4dc6-b8cb-8ea323977636 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.848386] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1201.848386] env[62814]: value = "task-4294711" [ 1201.848386] env[62814]: _type = "Task" [ 1201.848386] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.856973] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294711, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.933268] env[62814]: DEBUG oslo_concurrency.lockutils [req-3ae92f24-a00c-48eb-aa3f-052e538c7629 req-426bbe89-b823-4412-a564-19788993fb9d service nova] Releasing lock "refresh_cache-0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1201.933619] env[62814]: DEBUG nova.compute.manager [req-3ae92f24-a00c-48eb-aa3f-052e538c7629 req-426bbe89-b823-4412-a564-19788993fb9d service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received event network-vif-deleted-fec67180-c78f-4b6c-b106-a0d1ab127ec4 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1201.973755] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Getting list of instances from cluster (obj){ [ 1201.973755] env[62814]: value = "domain-c8" [ 1201.973755] env[62814]: _type = "ClusterComputeResource" [ 1201.973755] env[62814]: } {{(pid=62814) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1201.974957] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a51b728-8da1-4dba-8114-a226493e71cc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.998788] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Got total of 8 instances {{(pid=62814) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1201.998956] env[62814]: WARNING nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] While synchronizing instance power states, found 11 instances in the database and 8 instances on the hypervisor. [ 1201.999097] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid 7350d352-9336-40b8-81a6-0a4795d9f8dd {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1201.999295] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid edcdb1e5-ed9d-49e6-97e6-ea7629682547 {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1201.999452] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid 1f13051e-e51b-4981-9445-d5420c3c1818 {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1201.999605] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid 2d7a18f3-456b-470c-a759-632e5d79a8f0 {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1201.999758] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid b1d88997-e52f-41bd-b1b4-dd096d20d60a {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1201.999905] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid 33ab2d9e-e960-40ae-acf9-4fea0d9c830e {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1202.000063] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid 606e31eb-2349-427f-9c9b-ed9dc5b385f5 {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1202.000214] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid 40e3c00b-2129-476e-bc67-a0ef13781990 {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1202.000359] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid bf2ba498-7c47-4985-a84a-475037deec55 {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1202.000503] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid 44d94dc4-d7b3-4799-9044-9c1ecc80c88c {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1202.000644] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Triggering sync for uuid 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac {{(pid=62814) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10975}} [ 1202.001018] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "7350d352-9336-40b8-81a6-0a4795d9f8dd" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.001236] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "7350d352-9336-40b8-81a6-0a4795d9f8dd" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1202.001503] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.001716] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "1f13051e-e51b-4981-9445-d5420c3c1818" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.001912] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.002121] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.002300] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1202.002467] env[62814]: INFO nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] During sync_power_state the instance has a pending task (resize_prep). Skip. [ 1202.002625] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1202.002807] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.003013] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.003190] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1202.003344] env[62814]: INFO nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] During sync_power_state the instance has a pending task (resize_prep). Skip. [ 1202.003497] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1202.003675] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "40e3c00b-2129-476e-bc67-a0ef13781990" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.003892] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "bf2ba498-7c47-4985-a84a-475037deec55" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.004108] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.004307] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1202.004481] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1202.004611] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 1202.005711] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ff3b984-044e-4f3f-846d-90bc3765b384 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.071930] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294707, 'name': CreateVM_Task, 'duration_secs': 1.110997} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.072130] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1202.072817] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.072985] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1202.073329] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1202.073629] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63419be4-a10d-4345-863b-634297a4ad1e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.087045] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294705, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.087624] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1202.087624] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b105ae-31e2-e24a-47e5-56b8aaf5ca8a" [ 1202.087624] env[62814]: _type = "Task" [ 1202.087624] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.093859] env[62814]: DEBUG oslo_vmware.api [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294710, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.098813] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52b105ae-31e2-e24a-47e5-56b8aaf5ca8a, 'name': SearchDatastore_Task, 'duration_secs': 0.012844} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.099102] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1202.099326] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1202.099556] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.099699] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1202.099870] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1202.100130] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8f0d665-142b-4403-ba9c-7357000c24b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.109124] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1202.109307] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1202.110009] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afe29bad-2404-4e5e-a4f7-d1b32de7d801 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.117429] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1202.117616] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1202.117788] env[62814]: DEBUG nova.network.neutron [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1202.119232] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1202.119232] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521e4d41-ac8d-cd28-6f18-4c83deb0b8ec" [ 1202.119232] env[62814]: _type = "Task" [ 1202.119232] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.127032] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521e4d41-ac8d-cd28-6f18-4c83deb0b8ec, 'name': SearchDatastore_Task, 'duration_secs': 0.008438} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.131810] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17d8d6e0-f1c3-4a81-b8e8-a5c014dcfaac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.141048] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1202.141048] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52083bc3-1121-d349-333d-477d1cbffa9a" [ 1202.141048] env[62814]: _type = "Task" [ 1202.141048] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.150665] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52083bc3-1121-d349-333d-477d1cbffa9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.270990] env[62814]: DEBUG nova.network.neutron [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance_info_cache with network_info: [{"id": "580abcb7-b7f2-437a-8a84-738e83af0751", "address": "fa:16:3e:f1:6c:e8", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap580abcb7-b7", "ovs_interfaceid": "580abcb7-b7f2-437a-8a84-738e83af0751", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1202.290463] env[62814]: DEBUG oslo_vmware.api [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294708, 'name': ReconfigVM_Task, 'duration_secs': 0.994267} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.293029] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Reconfigured VM instance instance-00000068 to attach disk [datastore1] volume-ea4badfa-3567-42a6-83c7-13ecccee11d3/volume-ea4badfa-3567-42a6-83c7-13ecccee11d3.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.298208] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-feaabcfc-416a-4d5e-96ce-cf9e7436e153 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.310568] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ca9083-4155-4587-b756-5693e8faa9b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.320733] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0091f2f1-6d03-4dc9-af91-a6f9b178109e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.323992] env[62814]: DEBUG oslo_vmware.api [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1202.323992] env[62814]: value = "task-4294712" [ 1202.323992] env[62814]: _type = "Task" [ 1202.323992] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.355493] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb684d6-989c-4b42-b285-69a46151d535 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.361598] env[62814]: DEBUG oslo_vmware.api [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294712, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.366515] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294711, 'name': ReconfigVM_Task, 'duration_secs': 0.38058} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.368715] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 1f13051e-e51b-4981-9445-d5420c3c1818/1f13051e-e51b-4981-9445-d5420c3c1818.vmdk or device None with type streamOptimized {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.370183] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_secret_uuid': None, 'size': 0, 'device_type': 'disk', 'device_name': '/dev/sda', 'boot_index': 0, 'encrypted': False, 'guest_format': None, 'encryption_format': None, 'disk_bus': None, 'encryption_options': None, 'image_id': 'e7d68632-de75-4206-8f75-4abd879e1d22'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '5d72190a-51a7-4c35-afae-c61d93fc3655', 'mount_device': '/dev/sdb', 'device_type': None, 'delete_on_termination': False, 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845855', 'volume_id': '3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'name': 'volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '1f13051e-e51b-4981-9445-d5420c3c1818', 'attached_at': '', 'detached_at': '', 'volume_id': '3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'serial': '3e53f486-225a-44b7-bf8a-b91c6ac86fda'}, 'guest_format': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=62814) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1202.370455] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1202.370691] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845855', 'volume_id': '3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'name': 'volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '1f13051e-e51b-4981-9445-d5420c3c1818', 'attached_at': '', 'detached_at': '', 'volume_id': '3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'serial': '3e53f486-225a-44b7-bf8a-b91c6ac86fda'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1202.371491] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1ce05d-4000-457b-a8e7-66131bf1afdb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.374948] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936ab37a-07fa-49b5-8966-f19edc9f1789 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.399348] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604dd192-270f-4400-a07c-abe42ba0ea62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.401508] env[62814]: DEBUG nova.compute.provider_tree [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1202.425874] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda/volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1202.426721] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1f45c8f-b26e-4e80-9057-b3dfaede02f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.444380] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1202.444380] env[62814]: value = "task-4294713" [ 1202.444380] env[62814]: _type = "Task" [ 1202.444380] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.452017] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294713, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.514327] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "7350d352-9336-40b8-81a6-0a4795d9f8dd" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.513s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1202.561184] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea5fdd4-f697-49a7-8aec-1c0e2ecb8ef4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.583137] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance 'b1d88997-e52f-41bd-b1b4-dd096d20d60a' progress to 0 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1202.597161] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294705, 'name': ReconfigVM_Task, 'duration_secs': 2.019913} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.600101] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 44d94dc4-d7b3-4799-9044-9c1ecc80c88c/44d94dc4-d7b3-4799-9044-9c1ecc80c88c.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.600758] env[62814]: DEBUG oslo_vmware.api [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294710, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.544604} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.600982] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cc423dc-c067-4f96-844c-47a65c13ad04 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.602455] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1202.602636] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1202.602855] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1202.603309] env[62814]: INFO nova.compute.manager [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Took 2.38 seconds to destroy the instance on the hypervisor. [ 1202.603309] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1202.604220] env[62814]: DEBUG nova.compute.manager [-] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1202.604220] env[62814]: DEBUG nova.network.neutron [-] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1202.610037] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1202.610037] env[62814]: value = "task-4294714" [ 1202.610037] env[62814]: _type = "Task" [ 1202.610037] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.618012] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294714, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.650067] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52083bc3-1121-d349-333d-477d1cbffa9a, 'name': SearchDatastore_Task, 'duration_secs': 0.009716} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.652385] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1202.652701] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac/0fdb57c1-7a2a-455d-acb8-9f342ef1dbac.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1202.652988] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5dfc9fa3-6978-4dc0-b0fd-97349e1b6f5f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.659277] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1202.659277] env[62814]: value = "task-4294715" [ 1202.659277] env[62814]: _type = "Task" [ 1202.659277] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.666627] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294715, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.673180] env[62814]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port fec67180-c78f-4b6c-b106-a0d1ab127ec4 could not be found.", "detail": ""}} {{(pid=62814) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 1202.673383] env[62814]: DEBUG nova.network.neutron [-] Unable to show port fec67180-c78f-4b6c-b106-a0d1ab127ec4 as it no longer exists. {{(pid=62814) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 1202.773515] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-40e3c00b-2129-476e-bc67-a0ef13781990" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1202.773725] env[62814]: DEBUG nova.objects.instance [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lazy-loading 'migration_context' on Instance uuid 40e3c00b-2129-476e-bc67-a0ef13781990 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1202.839256] env[62814]: DEBUG oslo_vmware.api [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294712, 'name': ReconfigVM_Task, 'duration_secs': 0.250118} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.839256] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845862', 'volume_id': 'ea4badfa-3567-42a6-83c7-13ecccee11d3', 'name': 'volume-ea4badfa-3567-42a6-83c7-13ecccee11d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '33ab2d9e-e960-40ae-acf9-4fea0d9c830e', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea4badfa-3567-42a6-83c7-13ecccee11d3', 'serial': 'ea4badfa-3567-42a6-83c7-13ecccee11d3'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1202.911228] env[62814]: DEBUG nova.scheduler.client.report [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1202.935132] env[62814]: DEBUG nova.network.neutron [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance_info_cache with network_info: [{"id": "837d03b5-250f-44d2-bafb-31184267117d", "address": "fa:16:3e:24:ee:bf", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap837d03b5-25", "ovs_interfaceid": "837d03b5-250f-44d2-bafb-31184267117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1202.956867] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294713, 'name': ReconfigVM_Task, 'duration_secs': 0.375841} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.957664] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda/volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1202.964547] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2de0901-7f75-4bf9-84ed-1e8909a39aa4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1202.982028] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1202.982028] env[62814]: value = "task-4294716" [ 1202.982028] env[62814]: _type = "Task" [ 1202.982028] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1202.991281] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294716, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.096025] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1203.096025] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8ef962a0-4d3c-436f-a92a-8857d14b8321 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.103901] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1203.103901] env[62814]: value = "task-4294717" [ 1203.103901] env[62814]: _type = "Task" [ 1203.103901] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.113048] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294717, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.121131] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294714, 'name': Rename_Task, 'duration_secs': 0.196858} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.121473] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1203.121777] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a124aa1e-2185-493a-8fe3-8deb1f3c7ee7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.128994] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1203.128994] env[62814]: value = "task-4294718" [ 1203.128994] env[62814]: _type = "Task" [ 1203.128994] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.137572] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294718, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.168897] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294715, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.277034] env[62814]: DEBUG nova.objects.base [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Object Instance<40e3c00b-2129-476e-bc67-a0ef13781990> lazy-loaded attributes: info_cache,migration_context {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1203.278208] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-796aef68-c6f1-4cdc-b9cd-4a5171eb066b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.299572] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a50ba143-2107-4552-bb02-1b8d8d197932 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.307272] env[62814]: DEBUG oslo_vmware.api [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1203.307272] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c37d37-5211-94a1-a101-9e6aa7fedc84" [ 1203.307272] env[62814]: _type = "Task" [ 1203.307272] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.315936] env[62814]: DEBUG oslo_vmware.api [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c37d37-5211-94a1-a101-9e6aa7fedc84, 'name': SearchDatastore_Task, 'duration_secs': 0.007131} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.316690] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1203.419639] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.816s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.422979] env[62814]: DEBUG oslo_concurrency.lockutils [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.393s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.423335] env[62814]: DEBUG nova.objects.instance [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lazy-loading 'resources' on Instance uuid edcdb1e5-ed9d-49e6-97e6-ea7629682547 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1203.437252] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1203.449179] env[62814]: INFO nova.scheduler.client.report [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted allocations for instance bf2ba498-7c47-4985-a84a-475037deec55 [ 1203.491873] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294716, 'name': ReconfigVM_Task, 'duration_secs': 0.340161} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.492263] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845855', 'volume_id': '3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'name': 'volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '1f13051e-e51b-4981-9445-d5420c3c1818', 'attached_at': '', 'detached_at': '', 'volume_id': '3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'serial': '3e53f486-225a-44b7-bf8a-b91c6ac86fda'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1203.492864] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e5d34fc-e742-4db0-b79c-0b26361085ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.500286] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1203.500286] env[62814]: value = "task-4294719" [ 1203.500286] env[62814]: _type = "Task" [ 1203.500286] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.510686] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294719, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.536780] env[62814]: DEBUG nova.compute.manager [req-0eda912e-93c6-471c-9af3-43adb5b2e1d2 req-9e8e8f1c-c72f-40ee-8673-4ad9f1e6de8f service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Received event network-vif-deleted-1c6b72d7-d148-4280-981c-9bfbf768bd3b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1203.536935] env[62814]: INFO nova.compute.manager [req-0eda912e-93c6-471c-9af3-43adb5b2e1d2 req-9e8e8f1c-c72f-40ee-8673-4ad9f1e6de8f service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Neutron deleted interface 1c6b72d7-d148-4280-981c-9bfbf768bd3b; detaching it from the instance and deleting it from the info cache [ 1203.537125] env[62814]: DEBUG nova.network.neutron [req-0eda912e-93c6-471c-9af3-43adb5b2e1d2 req-9e8e8f1c-c72f-40ee-8673-4ad9f1e6de8f service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1203.615251] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294717, 'name': PowerOffVM_Task, 'duration_secs': 0.356974} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.616255] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1203.616737] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance 'b1d88997-e52f-41bd-b1b4-dd096d20d60a' progress to 17 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1203.638420] env[62814]: DEBUG oslo_vmware.api [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294718, 'name': PowerOnVM_Task, 'duration_secs': 0.480007} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.638787] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1203.639074] env[62814]: INFO nova.compute.manager [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Took 9.98 seconds to spawn the instance on the hypervisor. [ 1203.639326] env[62814]: DEBUG nova.compute.manager [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1203.640188] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847a3575-7a74-4a57-9bb8-22090124e70b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.669977] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294715, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546848} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1203.670436] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac/0fdb57c1-7a2a-455d-acb8-9f342ef1dbac.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1203.670654] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1203.670950] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6123d219-f88b-4311-a015-f31f5a985148 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.677114] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1203.677114] env[62814]: value = "task-4294720" [ 1203.677114] env[62814]: _type = "Task" [ 1203.677114] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1203.686585] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294720, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1203.870824] env[62814]: DEBUG nova.network.neutron [-] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1203.901420] env[62814]: DEBUG nova.objects.instance [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'flavor' on Instance uuid 33ab2d9e-e960-40ae-acf9-4fea0d9c830e {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1203.956822] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1ef081da-eec3-414f-acaa-6f5de3d770af tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "bf2ba498-7c47-4985-a84a-475037deec55" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.525s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1203.958163] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "bf2ba498-7c47-4985-a84a-475037deec55" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.954s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1203.958622] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b75e7327-64f9-45a6-99aa-eb17375cad96 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.970683] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2988c471-893d-4665-8f28-8c86cabadf33 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.018910] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294719, 'name': Rename_Task, 'duration_secs': 0.226595} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.019774] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1204.020037] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd130808-d7c2-439d-9888-b2f807a484ad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.029149] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1204.029149] env[62814]: value = "task-4294721" [ 1204.029149] env[62814]: _type = "Task" [ 1204.029149] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.035889] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294721, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.042360] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eba7f001-8b62-4626-b6ea-5b2fd3834df5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.050570] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a773f3f-b13a-46ed-b247-3f4658040419 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.083424] env[62814]: DEBUG nova.compute.manager [req-0eda912e-93c6-471c-9af3-43adb5b2e1d2 req-9e8e8f1c-c72f-40ee-8673-4ad9f1e6de8f service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Detach interface failed, port_id=1c6b72d7-d148-4280-981c-9bfbf768bd3b, reason: Instance 2d7a18f3-456b-470c-a759-632e5d79a8f0 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1204.127055] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1204.127908] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1204.127908] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1204.127908] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1204.127908] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1204.128164] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1204.128276] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1204.128449] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1204.128618] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1204.128790] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1204.129016] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1204.135838] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2482c4c-2b41-4e20-a040-5e877c184ad7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.159538] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1204.159538] env[62814]: value = "task-4294722" [ 1204.159538] env[62814]: _type = "Task" [ 1204.159538] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.168608] env[62814]: INFO nova.compute.manager [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Took 20.07 seconds to build instance. [ 1204.176691] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294722, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.187275] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294720, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.302137} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.190029] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1204.191680] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d908fc-7bc0-4f21-9ed0-846bb99b387d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.213680] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac/0fdb57c1-7a2a-455d-acb8-9f342ef1dbac.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1204.215344] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0abbd4f-3de2-4e77-9ffb-50a978d4f746 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.230418] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5329432-89e5-464c-b9bd-d7961ffc4dd4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.242472] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d0c89f-b9af-457d-816f-81334f91c8ce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.245460] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1204.245460] env[62814]: value = "task-4294723" [ 1204.245460] env[62814]: _type = "Task" [ 1204.245460] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.280924] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398e1b79-924a-4e77-92ec-6bc17ab85393 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.286678] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294723, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.291334] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed2e197-dc96-46c8-b7c0-0777794bc3c4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.304688] env[62814]: DEBUG nova.compute.provider_tree [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1204.376286] env[62814]: INFO nova.compute.manager [-] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Took 1.77 seconds to deallocate network for instance. [ 1204.414244] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bce9d3de-31e7-4ff3-b3c9-ee6c02fe661a tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.858s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1204.415202] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.412s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1204.416095] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2d18b0-e6e5-48ae-8b8d-c5867d5ec544 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.515422] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "bf2ba498-7c47-4985-a84a-475037deec55" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.557s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1204.537730] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294721, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.653930] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d2b2fb-264e-4eda-899d-e4fe9a09e65d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.662860] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ab43f48-ea9a-4a0b-966e-5eff6e25803d tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Suspending the VM {{(pid=62814) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1204.666280] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-c30e8d0f-83e5-4b2d-9b2e-dc79facd34fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.673361] env[62814]: DEBUG oslo_concurrency.lockutils [None req-adca2991-1dca-4968-94dc-769ef6890a33 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.579s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1204.673639] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294722, 'name': ReconfigVM_Task, 'duration_secs': 0.265554} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.674978] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.671s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1204.675218] env[62814]: INFO nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] During sync_power_state the instance has a pending task (spawning). Skip. [ 1204.675431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1204.675710] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance 'b1d88997-e52f-41bd-b1b4-dd096d20d60a' progress to 33 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1204.679129] env[62814]: DEBUG oslo_vmware.api [None req-3ab43f48-ea9a-4a0b-966e-5eff6e25803d tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1204.679129] env[62814]: value = "task-4294724" [ 1204.679129] env[62814]: _type = "Task" [ 1204.679129] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.687122] env[62814]: DEBUG oslo_vmware.api [None req-3ab43f48-ea9a-4a0b-966e-5eff6e25803d tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294724, 'name': SuspendVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.755383] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294723, 'name': ReconfigVM_Task, 'duration_secs': 0.349485} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1204.756386] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac/0fdb57c1-7a2a-455d-acb8-9f342ef1dbac.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1204.756386] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4823930-3a63-4a2a-bdaf-4a27065a9b08 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.761585] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1204.761585] env[62814]: value = "task-4294725" [ 1204.761585] env[62814]: _type = "Task" [ 1204.761585] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1204.770041] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294725, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1204.808526] env[62814]: DEBUG nova.scheduler.client.report [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1204.898238] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1204.932209] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.517s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1204.952340] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983f70e9-6e29-4456-a08e-f450ae85319a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1204.975730] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance '606e31eb-2349-427f-9c9b-ed9dc5b385f5' progress to 0 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1205.039144] env[62814]: DEBUG oslo_vmware.api [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294721, 'name': PowerOnVM_Task, 'duration_secs': 0.683619} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.039449] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1205.157426] env[62814]: DEBUG nova.compute.manager [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1205.158467] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14bf1c1-70a1-4769-9836-3748ea8b35b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.182045] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1205.182326] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1205.182455] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1205.182638] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1205.182783] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1205.182937] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1205.183140] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1205.183305] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1205.183467] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1205.183631] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1205.183855] env[62814]: DEBUG nova.virt.hardware [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1205.189123] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1205.192348] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87b7fa4c-a853-4b7a-852e-ba41a8c2d856 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.212885] env[62814]: DEBUG oslo_vmware.api [None req-3ab43f48-ea9a-4a0b-966e-5eff6e25803d tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294724, 'name': SuspendVM_Task} progress is 62%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.214237] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1205.214237] env[62814]: value = "task-4294726" [ 1205.214237] env[62814]: _type = "Task" [ 1205.214237] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.222411] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294726, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.271659] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294725, 'name': Rename_Task, 'duration_secs': 0.230101} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.271991] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1205.272324] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da3dd519-e6fb-4f48-a4f0-5f371ada40ad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.277588] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1205.277588] env[62814]: value = "task-4294727" [ 1205.277588] env[62814]: _type = "Task" [ 1205.277588] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.285146] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294727, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.314292] env[62814]: DEBUG oslo_concurrency.lockutils [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.891s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.316694] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.339390] env[62814]: INFO nova.scheduler.client.report [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Deleted allocations for instance edcdb1e5-ed9d-49e6-97e6-ea7629682547 [ 1205.429625] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1205.429876] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.482524] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1205.482932] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46e3a2bf-708a-4fc7-aa31-b6b43a0d1aa7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.489431] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1205.489431] env[62814]: value = "task-4294728" [ 1205.489431] env[62814]: _type = "Task" [ 1205.489431] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.500194] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294728, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.597646] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "c9b67326-3058-4fa6-a094-b47bf8663444" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1205.597948] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "c9b67326-3058-4fa6-a094-b47bf8663444" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.677362] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7db2d2f3-e3ce-4553-b25c-775c264bc38b tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 29.233s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.678204] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 3.676s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.678402] env[62814]: INFO nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] During sync_power_state the instance has a pending task (spawning). Skip. [ 1205.678578] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.688709] env[62814]: DEBUG oslo_vmware.api [None req-3ab43f48-ea9a-4a0b-966e-5eff6e25803d tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294724, 'name': SuspendVM_Task, 'duration_secs': 0.855215} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.689036] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ab43f48-ea9a-4a0b-966e-5eff6e25803d tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Suspended the VM {{(pid=62814) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1205.689319] env[62814]: DEBUG nova.compute.manager [None req-3ab43f48-ea9a-4a0b-966e-5eff6e25803d tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1205.690238] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546240a9-82fb-4fe4-9b4b-fbaab5acc141 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.723954] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294726, 'name': ReconfigVM_Task, 'duration_secs': 0.180281} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.724293] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1205.725109] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f5be01-3be3-4eef-9239-6932f1c60935 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.751151] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] b1d88997-e52f-41bd-b1b4-dd096d20d60a/b1d88997-e52f-41bd-b1b4-dd096d20d60a.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1205.751694] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9acf035e-6f7d-4d20-a48e-9133ff7c23c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.770279] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1205.770279] env[62814]: value = "task-4294729" [ 1205.770279] env[62814]: _type = "Task" [ 1205.770279] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.778353] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294729, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.786234] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294727, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.847938] env[62814]: DEBUG oslo_concurrency.lockutils [None req-778b16e1-298c-42e7-a16f-f8d5b3f5cecb tempest-ServersNegativeTestJSON-1201522932 tempest-ServersNegativeTestJSON-1201522932-project-member] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.724s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1205.848877] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 3.847s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1205.849362] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc5d9a0a-239c-4f77-883f-04f01b6d8c15 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.859956] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1096f093-9f1d-4835-81da-9f3ea7f714b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.932914] env[62814]: INFO nova.compute.manager [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Detaching volume c2969fdc-ffa5-4fde-b88c-26966e10e336 [ 1205.971671] env[62814]: INFO nova.virt.block_device [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Attempting to driver detach volume c2969fdc-ffa5-4fde-b88c-26966e10e336 from mountpoint /dev/sdb [ 1205.971671] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1205.971671] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845858', 'volume_id': 'c2969fdc-ffa5-4fde-b88c-26966e10e336', 'name': 'volume-c2969fdc-ffa5-4fde-b88c-26966e10e336', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '33ab2d9e-e960-40ae-acf9-4fea0d9c830e', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2969fdc-ffa5-4fde-b88c-26966e10e336', 'serial': 'c2969fdc-ffa5-4fde-b88c-26966e10e336'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1205.972572] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926f8f45-916d-42b6-ba61-35a38b09939a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.007526] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa4545c-e506-47b3-81b5-0cc59eef6edb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.015489] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294728, 'name': PowerOffVM_Task, 'duration_secs': 0.208704} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.016975] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1206.017193] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance '606e31eb-2349-427f-9c9b-ed9dc5b385f5' progress to 17 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1206.023591] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e444919d-5ca6-436d-bca6-0104333c6cfd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.048384] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1430f8e-5751-488e-88b3-0f542bfc4ac4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.051669] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd691a7b-d904-4f44-8732-7d591f5940fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.071811] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3414ca-6de3-4b9a-9d80-df2affbe60c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.076199] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] The volume has not been displaced from its original location: [datastore1] volume-c2969fdc-ffa5-4fde-b88c-26966e10e336/volume-c2969fdc-ffa5-4fde-b88c-26966e10e336.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1206.082682] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Reconfiguring VM instance instance-00000068 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1206.082682] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bcb7e14-1a9e-4799-9061-d2f4fd3ec094 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.122901] env[62814]: DEBUG nova.compute.manager [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1206.129208] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb4656f-754a-401f-a7ad-c1f3cc1384d6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.131221] env[62814]: DEBUG oslo_vmware.api [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1206.131221] env[62814]: value = "task-4294730" [ 1206.131221] env[62814]: _type = "Task" [ 1206.131221] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.138399] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31484709-3f27-4632-ab85-21e59e3daa03 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.145185] env[62814]: DEBUG oslo_vmware.api [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294730, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.155012] env[62814]: DEBUG nova.compute.provider_tree [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1206.282292] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294729, 'name': ReconfigVM_Task, 'duration_secs': 0.295843} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.285178] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfigured VM instance instance-00000067 to attach disk [datastore2] b1d88997-e52f-41bd-b1b4-dd096d20d60a/b1d88997-e52f-41bd-b1b4-dd096d20d60a.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1206.285457] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance 'b1d88997-e52f-41bd-b1b4-dd096d20d60a' progress to 50 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1206.293666] env[62814]: DEBUG oslo_vmware.api [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294727, 'name': PowerOnVM_Task, 'duration_secs': 0.826808} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.293925] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1206.294151] env[62814]: INFO nova.compute.manager [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Took 8.42 seconds to spawn the instance on the hypervisor. [ 1206.294332] env[62814]: DEBUG nova.compute.manager [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1206.295111] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e785a3-7cf9-4dd7-8a99-186d3c643c26 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.402540] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "edcdb1e5-ed9d-49e6-97e6-ea7629682547" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.553s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1206.529162] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1206.529457] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1206.529589] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1206.529772] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1206.529915] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1206.530075] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1206.530314] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1206.530464] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1206.530629] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1206.530799] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1206.530978] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1206.537300] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a8b2ff0-fcaf-4c46-a8dc-eea35c1153fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.556541] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1206.556541] env[62814]: value = "task-4294731" [ 1206.556541] env[62814]: _type = "Task" [ 1206.556541] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.565873] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294731, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.642156] env[62814]: DEBUG oslo_vmware.api [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294730, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.647746] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1206.657878] env[62814]: DEBUG nova.scheduler.client.report [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1206.793023] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e1bdef4-d471-4f50-ae8e-d8d6a13f52db {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.821782] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ec9240-1d1c-472b-87d2-5782691b8aae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.824719] env[62814]: INFO nova.compute.manager [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Took 18.02 seconds to build instance. [ 1206.844471] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance 'b1d88997-e52f-41bd-b1b4-dd096d20d60a' progress to 67 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1207.052102] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1207.052465] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1207.052689] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1207.052874] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1207.053091] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.056397] env[62814]: INFO nova.compute.manager [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Terminating instance [ 1207.068456] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294731, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.144087] env[62814]: DEBUG oslo_vmware.api [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294730, 'name': ReconfigVM_Task, 'duration_secs': 0.554717} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.144425] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Reconfigured VM instance instance-00000068 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1207.149417] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ca62a8c-5e3c-43d4-a18e-d80ac97a20a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.168269] env[62814]: DEBUG oslo_vmware.api [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1207.168269] env[62814]: value = "task-4294732" [ 1207.168269] env[62814]: _type = "Task" [ 1207.168269] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.179655] env[62814]: DEBUG oslo_vmware.api [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294732, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.327228] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5853a357-d596-4cda-8518-ef834df34338 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.538s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.327572] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 5.323s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1207.327839] env[62814]: INFO nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] During sync_power_state the instance has a pending task (spawning). Skip. [ 1207.328134] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.567373] env[62814]: DEBUG nova.compute.manager [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1207.567652] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1207.569407] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ed95e2-4e7e-4924-9a25-bd4b2c2aa101 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.579324] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294731, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.582445] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1207.582976] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-86df279d-803d-4f44-967b-ac2993d1b785 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.642680] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1207.642922] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1207.643131] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleting the datastore file [datastore2] 44d94dc4-d7b3-4799-9044-9c1ecc80c88c {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1207.643397] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-592ebb30-685e-41b9-98c9-b6a236be5161 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.652603] env[62814]: DEBUG oslo_vmware.api [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1207.652603] env[62814]: value = "task-4294734" [ 1207.652603] env[62814]: _type = "Task" [ 1207.652603] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.658765] env[62814]: DEBUG oslo_vmware.api [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294734, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.671494] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.353s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1207.673086] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.775s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1207.673334] env[62814]: DEBUG nova.objects.instance [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'resources' on Instance uuid 2d7a18f3-456b-470c-a759-632e5d79a8f0 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1207.684778] env[62814]: DEBUG oslo_vmware.api [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294732, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.069717] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294731, 'name': ReconfigVM_Task, 'duration_secs': 1.230184} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.070699] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance '606e31eb-2349-427f-9c9b-ed9dc5b385f5' progress to 33 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1208.161917] env[62814]: DEBUG oslo_vmware.api [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294734, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.315228} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.162280] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1208.162784] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1208.165518] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1208.165518] env[62814]: INFO nova.compute.manager [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1208.165518] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1208.165518] env[62814]: DEBUG nova.compute.manager [-] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1208.165518] env[62814]: DEBUG nova.network.neutron [-] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1208.189325] env[62814]: DEBUG oslo_vmware.api [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294732, 'name': ReconfigVM_Task, 'duration_secs': 0.723467} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.190367] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845858', 'volume_id': 'c2969fdc-ffa5-4fde-b88c-26966e10e336', 'name': 'volume-c2969fdc-ffa5-4fde-b88c-26966e10e336', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '33ab2d9e-e960-40ae-acf9-4fea0d9c830e', 'attached_at': '', 'detached_at': '', 'volume_id': 'c2969fdc-ffa5-4fde-b88c-26966e10e336', 'serial': 'c2969fdc-ffa5-4fde-b88c-26966e10e336'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1208.261060] env[62814]: INFO nova.scheduler.client.report [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted allocation for migration c7cb01ac-d4df-41dd-b19a-6835a05449a3 [ 1208.411260] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d987e9ac-8c31-4cc2-8e41-4d6937f7cbcd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.420021] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3008999-48e1-4524-a989-53ff0f321fb4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.429020] env[62814]: DEBUG nova.compute.manager [req-30fef480-b06b-4bdd-b9fb-52171485f9ae req-f0a531f4-614c-4294-accf-dc1a015d253a service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Received event network-changed-e07ce73b-c8ba-4a79-a9a4-a5d1672041d9 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1208.429020] env[62814]: DEBUG nova.compute.manager [req-30fef480-b06b-4bdd-b9fb-52171485f9ae req-f0a531f4-614c-4294-accf-dc1a015d253a service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Refreshing instance network info cache due to event network-changed-e07ce73b-c8ba-4a79-a9a4-a5d1672041d9. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1208.429020] env[62814]: DEBUG oslo_concurrency.lockutils [req-30fef480-b06b-4bdd-b9fb-52171485f9ae req-f0a531f4-614c-4294-accf-dc1a015d253a service nova] Acquiring lock "refresh_cache-0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1208.429020] env[62814]: DEBUG oslo_concurrency.lockutils [req-30fef480-b06b-4bdd-b9fb-52171485f9ae req-f0a531f4-614c-4294-accf-dc1a015d253a service nova] Acquired lock "refresh_cache-0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1208.429020] env[62814]: DEBUG nova.network.neutron [req-30fef480-b06b-4bdd-b9fb-52171485f9ae req-f0a531f4-614c-4294-accf-dc1a015d253a service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Refreshing network info cache for port e07ce73b-c8ba-4a79-a9a4-a5d1672041d9 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1209.196315] env[62814]: DEBUG nova.objects.instance [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'flavor' on Instance uuid 33ab2d9e-e960-40ae-acf9-4fea0d9c830e {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1209.200214] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1209.200510] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1209.200603] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1209.200769] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1209.200919] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1209.201540] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1209.201540] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1209.201540] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1209.201540] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1209.201839] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1209.201947] env[62814]: DEBUG nova.virt.hardware [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1209.207159] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1209.208039] env[62814]: DEBUG oslo_concurrency.lockutils [None req-51ae1091-dd73-4be0-838d-6cc791c281e0 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "40e3c00b-2129-476e-bc67-a0ef13781990" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.771s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1209.212394] env[62814]: DEBUG nova.compute.manager [req-cc6a2fb5-f42e-4073-a358-8ab37348aca4 req-280d66a3-3e57-49cd-9a1c-fb046af8798c service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Received event network-vif-deleted-559f235e-f765-465d-a9be-74bb8d7ed03b {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1209.212394] env[62814]: INFO nova.compute.manager [req-cc6a2fb5-f42e-4073-a358-8ab37348aca4 req-280d66a3-3e57-49cd-9a1c-fb046af8798c service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Neutron deleted interface 559f235e-f765-465d-a9be-74bb8d7ed03b; detaching it from the instance and deleting it from the info cache [ 1209.212394] env[62814]: DEBUG nova.network.neutron [req-cc6a2fb5-f42e-4073-a358-8ab37348aca4 req-280d66a3-3e57-49cd-9a1c-fb046af8798c service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.212394] env[62814]: DEBUG nova.network.neutron [-] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.212853] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12bf48a9-f85d-48f0-bd2a-cc0db5a5d908 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.225502] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "40e3c00b-2129-476e-bc67-a0ef13781990" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.222s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1209.226307] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129dcf65-db03-49d2-abeb-7da4d12c449b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.230755] env[62814]: INFO nova.compute.manager [-] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Took 1.07 seconds to deallocate network for instance. [ 1209.231962] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ad8d2d-d395-4188-8e1a-42a2f35fdd25 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.245206] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcc5fa00-607d-47f1-8ba6-9c211acf06f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.254288] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1209.254288] env[62814]: value = "task-4294735" [ 1209.254288] env[62814]: _type = "Task" [ 1209.254288] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.268610] env[62814]: DEBUG nova.compute.provider_tree [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1209.272328] env[62814]: DEBUG nova.network.neutron [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Port 0235e89b-ec0d-4439-817c-dd76d6987b80 binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1209.277153] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294735, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.480768] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "fed47b28-64fb-4af9-9f8e-97a63afda514" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1209.480964] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "fed47b28-64fb-4af9-9f8e-97a63afda514" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1209.481967] env[62814]: WARNING oslo_messaging._drivers.amqpdriver [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 1209.697608] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "40e3c00b-2129-476e-bc67-a0ef13781990" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1209.736998] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8354689a-3047-4543-b87a-51601a5de6d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.743932] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1209.747786] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3f7e16-e34a-46df-a1e0-85d781cddfd3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.758502] env[62814]: INFO nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] During sync_power_state the instance has a pending task (deleting). Skip. [ 1209.758707] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "40e3c00b-2129-476e-bc67-a0ef13781990" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.533s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1209.759148] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1209.759404] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "40e3c00b-2129-476e-bc67-a0ef13781990" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.062s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1209.759596] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "40e3c00b-2129-476e-bc67-a0ef13781990-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1209.759775] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "40e3c00b-2129-476e-bc67-a0ef13781990-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1209.759933] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "40e3c00b-2129-476e-bc67-a0ef13781990-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1209.764920] env[62814]: INFO nova.compute.manager [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Terminating instance [ 1209.771342] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294735, 'name': ReconfigVM_Task, 'duration_secs': 0.402742} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.771590] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1209.772985] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f03c3178-0b40-42f8-aa8f-b2d83c023a8d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.787494] env[62814]: DEBUG nova.scheduler.client.report [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1209.790412] env[62814]: DEBUG nova.compute.manager [req-cc6a2fb5-f42e-4073-a358-8ab37348aca4 req-280d66a3-3e57-49cd-9a1c-fb046af8798c service nova] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Detach interface failed, port_id=559f235e-f765-465d-a9be-74bb8d7ed03b, reason: Instance 44d94dc4-d7b3-4799-9044-9c1ecc80c88c could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1209.813816] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 606e31eb-2349-427f-9c9b-ed9dc5b385f5/606e31eb-2349-427f-9c9b-ed9dc5b385f5.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1209.814409] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5ef8e3e6-a21f-4aed-9c3e-82ba44a35843 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.834990] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1209.834990] env[62814]: value = "task-4294736" [ 1209.834990] env[62814]: _type = "Task" [ 1209.834990] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.843784] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294736, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.938143] env[62814]: DEBUG nova.network.neutron [req-30fef480-b06b-4bdd-b9fb-52171485f9ae req-f0a531f4-614c-4294-accf-dc1a015d253a service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Updated VIF entry in instance network info cache for port e07ce73b-c8ba-4a79-a9a4-a5d1672041d9. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1209.938537] env[62814]: DEBUG nova.network.neutron [req-30fef480-b06b-4bdd-b9fb-52171485f9ae req-f0a531f4-614c-4294-accf-dc1a015d253a service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Updating instance_info_cache with network_info: [{"id": "e07ce73b-c8ba-4a79-a9a4-a5d1672041d9", "address": "fa:16:3e:86:21:ad", "network": {"id": "9fcd6c06-615a-4f9f-b3fd-240d4ee40859", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-515268172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.161", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "65a886f76f404ce5a6d0841241d8f120", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7e0240aa-a694-48fc-a0f9-6f2d3e71aa12", "external-id": "nsx-vlan-transportzone-249", "segmentation_id": 249, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape07ce73b-c8", "ovs_interfaceid": "e07ce73b-c8ba-4a79-a9a4-a5d1672041d9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.985360] env[62814]: DEBUG nova.compute.manager [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1210.218016] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4fefdeae-416d-4059-ba67-9a183f9eeb8d tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.788s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.219263] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.460s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.269387] env[62814]: DEBUG nova.compute.manager [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1210.269614] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1210.272651] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b500e42d-f14b-4400-aa79-454718d5587c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.280048] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1210.280290] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a91511c7-e13c-4b88-bd3c-a85affe0a7cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.286059] env[62814]: DEBUG oslo_vmware.api [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1210.286059] env[62814]: value = "task-4294737" [ 1210.286059] env[62814]: _type = "Task" [ 1210.286059] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.301906] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.629s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.303679] env[62814]: DEBUG oslo_vmware.api [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294737, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.307104] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.659s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.308544] env[62814]: INFO nova.compute.claims [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1210.314636] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1210.315335] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.315335] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.329050] env[62814]: INFO nova.scheduler.client.report [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleted allocations for instance 2d7a18f3-456b-470c-a759-632e5d79a8f0 [ 1210.345485] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294736, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.441332] env[62814]: DEBUG oslo_concurrency.lockutils [req-30fef480-b06b-4bdd-b9fb-52171485f9ae req-f0a531f4-614c-4294-accf-dc1a015d253a service nova] Releasing lock "refresh_cache-0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1210.504856] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1210.721882] env[62814]: INFO nova.compute.manager [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Detaching volume ea4badfa-3567-42a6-83c7-13ecccee11d3 [ 1210.757626] env[62814]: INFO nova.virt.block_device [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Attempting to driver detach volume ea4badfa-3567-42a6-83c7-13ecccee11d3 from mountpoint /dev/sdc [ 1210.757879] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1210.758156] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845862', 'volume_id': 'ea4badfa-3567-42a6-83c7-13ecccee11d3', 'name': 'volume-ea4badfa-3567-42a6-83c7-13ecccee11d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '33ab2d9e-e960-40ae-acf9-4fea0d9c830e', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea4badfa-3567-42a6-83c7-13ecccee11d3', 'serial': 'ea4badfa-3567-42a6-83c7-13ecccee11d3'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1210.760055] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cccd10d-ab92-49e8-8a84-607611e4727c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.780765] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85457df8-bf6c-4882-8b57-33666c667d2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.790281] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262a191b-77a3-4666-bd83-27ef446bdfe9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.796830] env[62814]: DEBUG oslo_vmware.api [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294737, 'name': PowerOffVM_Task, 'duration_secs': 0.245343} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.811911] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1210.811911] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1210.811911] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-522c9490-c0f0-4c18-a33e-f5b68fb7099f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.814445] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9e2491-36bf-4d35-8b2f-cf7c1e5eac12 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.831645] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] The volume has not been displaced from its original location: [datastore1] volume-ea4badfa-3567-42a6-83c7-13ecccee11d3/volume-ea4badfa-3567-42a6-83c7-13ecccee11d3.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1210.836792] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Reconfiguring VM instance instance-00000068 to detach disk 2002 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1210.839822] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e831c81-4687-4c1a-aec6-6f7ef7ba6af8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.853487] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8bc410fb-33a3-449c-9c05-211897304930 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.141s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1210.857074] env[62814]: DEBUG oslo_concurrency.lockutils [req-d71bca2c-6950-41a0-a3ac-bcbc41237ac8 req-943c67db-2894-4fd8-8770-f755c20939bc service nova] Acquired lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1210.858525] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5025bcfd-bb2c-4917-9749-f3e5c69f7039 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.862867] env[62814]: DEBUG oslo_vmware.api [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1210.862867] env[62814]: value = "task-4294739" [ 1210.862867] env[62814]: _type = "Task" [ 1210.862867] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.866101] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294736, 'name': ReconfigVM_Task, 'duration_secs': 0.630713} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.867154] env[62814]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1210.867302] env[62814]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=62814) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1210.870152] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 606e31eb-2349-427f-9c9b-ed9dc5b385f5/606e31eb-2349-427f-9c9b-ed9dc5b385f5.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1210.870424] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance '606e31eb-2349-427f-9c9b-ed9dc5b385f5' progress to 50 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1210.875097] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e45f37bd-ef9a-46bc-9f4a-6b4b5373a100 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.877159] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1210.877353] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1210.877527] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleting the datastore file [datastore2] 40e3c00b-2129-476e-bc67-a0ef13781990 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1210.878436] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9d536901-69b0-4340-955f-f6c44722b8e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.883166] env[62814]: DEBUG oslo_vmware.api [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294739, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.887994] env[62814]: DEBUG oslo_vmware.api [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1210.887994] env[62814]: value = "task-4294740" [ 1210.887994] env[62814]: _type = "Task" [ 1210.887994] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.895640] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa06117-fd13-4634-b5c2-cab8463a02f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.910329] env[62814]: DEBUG oslo_vmware.api [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294740, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.930803] env[62814]: ERROR root [req-d71bca2c-6950-41a0-a3ac-bcbc41237ac8 req-943c67db-2894-4fd8-8770-f755c20939bc service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-845832' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 479, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-845832' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-845832' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-845832'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-845832' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-845832' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-845832'}\n"]: nova.exception.InstanceNotFound: Instance 2d7a18f3-456b-470c-a759-632e5d79a8f0 could not be found. [ 1210.931075] env[62814]: DEBUG oslo_concurrency.lockutils [req-d71bca2c-6950-41a0-a3ac-bcbc41237ac8 req-943c67db-2894-4fd8-8770-f755c20939bc service nova] Releasing lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1210.931420] env[62814]: DEBUG nova.compute.manager [req-d71bca2c-6950-41a0-a3ac-bcbc41237ac8 req-943c67db-2894-4fd8-8770-f755c20939bc service nova] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Detach interface failed, port_id=88931aee-29b6-4f58-9ee6-0210612c8211, reason: Instance 2d7a18f3-456b-470c-a759-632e5d79a8f0 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1210.931862] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 8.930s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1210.932076] env[62814]: INFO nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] During sync_power_state the instance has a pending task (deleting). Skip. [ 1210.932251] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "2d7a18f3-456b-470c-a759-632e5d79a8f0" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1211.379829] env[62814]: DEBUG oslo_vmware.api [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294739, 'name': ReconfigVM_Task, 'duration_secs': 0.209445} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.380844] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f576ae7-6c76-404a-8c96-32f908d9d4b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.383431] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Reconfigured VM instance instance-00000068 to detach disk 2002 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1211.392557] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0426a9c6-80be-44d8-94f8-0493b15f6b91 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.402397] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.402397] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1211.402397] env[62814]: DEBUG nova.network.neutron [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1211.425827] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5988520b-5da7-4097-a0f4-c435dce62288 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.432242] env[62814]: DEBUG oslo_vmware.api [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1211.432242] env[62814]: value = "task-4294741" [ 1211.432242] env[62814]: _type = "Task" [ 1211.432242] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.432733] env[62814]: DEBUG oslo_vmware.api [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294740, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176349} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.433357] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1211.433542] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1211.433714] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1211.433901] env[62814]: INFO nova.compute.manager [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Took 1.16 seconds to destroy the instance on the hypervisor. [ 1211.434161] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1211.437410] env[62814]: DEBUG nova.compute.manager [-] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1211.437521] env[62814]: DEBUG nova.network.neutron [-] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1211.452264] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance '606e31eb-2349-427f-9c9b-ed9dc5b385f5' progress to 67 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1211.463760] env[62814]: DEBUG oslo_vmware.api [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294741, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.568244] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7698901-cf6a-4ddd-92b9-c114c01b1a61 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.579016] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e21653b-52e6-40ec-a305-ffa608be12fb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.609390] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b97ca57-5c9a-40be-8ccf-4149ca78cbf2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.616693] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ae9873-0b43-402b-b3b5-62dafbc9c578 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.629801] env[62814]: DEBUG nova.compute.provider_tree [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1211.816339] env[62814]: DEBUG nova.compute.manager [req-54afec48-749b-4824-b4c4-39d9930537f9 req-530f6924-110c-4a7b-a2d4-16042e5faa2d service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Received event network-vif-deleted-580abcb7-b7f2-437a-8a84-738e83af0751 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1211.816651] env[62814]: INFO nova.compute.manager [req-54afec48-749b-4824-b4c4-39d9930537f9 req-530f6924-110c-4a7b-a2d4-16042e5faa2d service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Neutron deleted interface 580abcb7-b7f2-437a-8a84-738e83af0751; detaching it from the instance and deleting it from the info cache [ 1211.816920] env[62814]: DEBUG nova.network.neutron [req-54afec48-749b-4824-b4c4-39d9930537f9 req-530f6924-110c-4a7b-a2d4-16042e5faa2d service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.943051] env[62814]: DEBUG oslo_vmware.api [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294741, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.000587] env[62814]: DEBUG nova.network.neutron [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Port 837d03b5-250f-44d2-bafb-31184267117d binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1212.154143] env[62814]: ERROR nova.scheduler.client.report [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [req-690a6c01-951c-4335-bbcc-cf1ee182d8b1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-690a6c01-951c-4335-bbcc-cf1ee182d8b1"}]} [ 1212.168826] env[62814]: DEBUG nova.network.neutron [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance_info_cache with network_info: [{"id": "0235e89b-ec0d-4439-817c-dd76d6987b80", "address": "fa:16:3e:46:9d:6d", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0235e89b-ec", "ovs_interfaceid": "0235e89b-ec0d-4439-817c-dd76d6987b80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.170659] env[62814]: DEBUG nova.scheduler.client.report [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1212.182240] env[62814]: DEBUG nova.scheduler.client.report [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1212.182446] env[62814]: DEBUG nova.compute.provider_tree [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1212.193313] env[62814]: DEBUG nova.scheduler.client.report [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1212.207902] env[62814]: DEBUG nova.scheduler.client.report [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1212.291502] env[62814]: DEBUG nova.network.neutron [-] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1212.320646] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68c7e6b8-2c80-45ab-b335-f7df2f6e199c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.330445] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03678057-5f34-4c7a-b479-abd8d0e2b658 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.361979] env[62814]: DEBUG nova.compute.manager [req-54afec48-749b-4824-b4c4-39d9930537f9 req-530f6924-110c-4a7b-a2d4-16042e5faa2d service nova] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Detach interface failed, port_id=580abcb7-b7f2-437a-8a84-738e83af0751, reason: Instance 40e3c00b-2129-476e-bc67-a0ef13781990 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1212.371080] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd75175-7f6e-4317-a445-c8db352d13de {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.377716] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4aa43c1-f7e9-4b38-b715-84a5d5e8953d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.408830] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6474ae5-c847-47cc-a669-fce7db35daf5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.417038] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2916be9-ff38-4a4b-9e6e-41e217efbf99 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.431706] env[62814]: DEBUG nova.compute.provider_tree [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1212.442734] env[62814]: DEBUG oslo_vmware.api [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294741, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.665384] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1212.665628] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1212.674225] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1212.794751] env[62814]: INFO nova.compute.manager [-] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Took 1.36 seconds to deallocate network for instance. [ 1212.945068] env[62814]: DEBUG oslo_vmware.api [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294741, 'name': ReconfigVM_Task, 'duration_secs': 1.21679} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.945406] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845862', 'volume_id': 'ea4badfa-3567-42a6-83c7-13ecccee11d3', 'name': 'volume-ea4badfa-3567-42a6-83c7-13ecccee11d3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '33ab2d9e-e960-40ae-acf9-4fea0d9c830e', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea4badfa-3567-42a6-83c7-13ecccee11d3', 'serial': 'ea4badfa-3567-42a6-83c7-13ecccee11d3'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1212.965404] env[62814]: DEBUG nova.scheduler.client.report [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 177 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1212.965653] env[62814]: DEBUG nova.compute.provider_tree [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 177 to 178 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1212.965829] env[62814]: DEBUG nova.compute.provider_tree [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1213.021566] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1213.021799] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1213.022179] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1213.168155] env[62814]: DEBUG nova.compute.manager [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1213.186435] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe1f982-aa3e-475d-9bcf-ec2161fd7b84 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.193861] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f67894-d917-4ce6-b3fd-047638cfdfa7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.302011] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1213.471050] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.164s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1213.471569] env[62814]: DEBUG nova.compute.manager [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1213.474390] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.731s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1213.474617] env[62814]: DEBUG nova.objects.instance [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lazy-loading 'resources' on Instance uuid 44d94dc4-d7b3-4799-9044-9c1ecc80c88c {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.591383] env[62814]: DEBUG nova.objects.instance [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'flavor' on Instance uuid 33ab2d9e-e960-40ae-acf9-4fea0d9c830e {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1213.703967] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1213.978176] env[62814]: DEBUG nova.compute.utils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1213.983636] env[62814]: DEBUG nova.compute.manager [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1213.984075] env[62814]: DEBUG nova.network.neutron [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1214.034368] env[62814]: DEBUG nova.policy [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd71116c0cfd4b509e4070ecc84613f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd82993ef7dfa4d2f8f39db3577dd321a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1214.071652] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1214.071863] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1214.072052] env[62814]: DEBUG nova.network.neutron [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1214.157380] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4c0f30-c533-4b18-8cd4-95bd3072fa48 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.165652] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf06dbe-3884-449b-b29c-d31fc2ffe028 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.200552] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8809b32a-e998-4b4c-88a1-98366c666ad7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.209233] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c0b200-f93e-4357-a90b-f5f74be35e41 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.224895] env[62814]: DEBUG nova.compute.provider_tree [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1214.288988] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d41eac21-1aef-4c0a-8d19-d600b8673928 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.310341] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09dac9a2-dfc3-45d1-8711-7281cdcea599 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.317243] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance 'b1d88997-e52f-41bd-b1b4-dd096d20d60a' progress to 83 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1214.333718] env[62814]: DEBUG nova.network.neutron [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Successfully created port: 382b237a-593e-47c2-99c0-18d3918d92a6 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1214.484683] env[62814]: DEBUG nova.compute.manager [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1214.598806] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e7856441-a08f-4ba3-85e6-a36035ab6773 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.379s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1214.730215] env[62814]: DEBUG nova.scheduler.client.report [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1214.824034] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1214.824403] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aca280b5-372f-424c-aca3-6345ce71a387 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1214.835464] env[62814]: DEBUG nova.network.neutron [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance_info_cache with network_info: [{"id": "837d03b5-250f-44d2-bafb-31184267117d", "address": "fa:16:3e:24:ee:bf", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap837d03b5-25", "ovs_interfaceid": "837d03b5-250f-44d2-bafb-31184267117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1214.842255] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1214.842255] env[62814]: value = "task-4294742" [ 1214.842255] env[62814]: _type = "Task" [ 1214.842255] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1214.851689] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294742, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.235649] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1215.241306] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.734s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1215.241306] env[62814]: INFO nova.compute.claims [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1215.263824] env[62814]: INFO nova.scheduler.client.report [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted allocations for instance 44d94dc4-d7b3-4799-9044-9c1ecc80c88c [ 1215.338789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1215.352526] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294742, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.495881] env[62814]: DEBUG nova.compute.manager [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1215.508836] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1215.509097] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1215.509318] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1215.509524] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1215.509706] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1215.515417] env[62814]: INFO nova.compute.manager [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Terminating instance [ 1215.520250] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1215.521037] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1215.521037] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1215.521037] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1215.521037] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1215.521337] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1215.521403] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1215.521542] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1215.521705] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1215.521878] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1215.522067] env[62814]: DEBUG nova.virt.hardware [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1215.522907] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc851a4-9583-4114-9804-f18eab9c716b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.532263] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26d0f382-26ca-4c45-aa4d-8a0b6726807f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.773440] env[62814]: DEBUG oslo_concurrency.lockutils [None req-2739c8c1-4753-493b-b208-d7917bed2f26 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "44d94dc4-d7b3-4799-9044-9c1ecc80c88c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.721s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1215.807091] env[62814]: DEBUG nova.compute.manager [req-8302993f-1d00-4dcb-bc60-3c0bff83546b req-fd3ddeb4-1374-493d-9385-e8cf86cd3ff9 service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Received event network-vif-plugged-382b237a-593e-47c2-99c0-18d3918d92a6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1215.807327] env[62814]: DEBUG oslo_concurrency.lockutils [req-8302993f-1d00-4dcb-bc60-3c0bff83546b req-fd3ddeb4-1374-493d-9385-e8cf86cd3ff9 service nova] Acquiring lock "c9b67326-3058-4fa6-a094-b47bf8663444-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1215.807535] env[62814]: DEBUG oslo_concurrency.lockutils [req-8302993f-1d00-4dcb-bc60-3c0bff83546b req-fd3ddeb4-1374-493d-9385-e8cf86cd3ff9 service nova] Lock "c9b67326-3058-4fa6-a094-b47bf8663444-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1215.808043] env[62814]: DEBUG oslo_concurrency.lockutils [req-8302993f-1d00-4dcb-bc60-3c0bff83546b req-fd3ddeb4-1374-493d-9385-e8cf86cd3ff9 service nova] Lock "c9b67326-3058-4fa6-a094-b47bf8663444-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1215.808043] env[62814]: DEBUG nova.compute.manager [req-8302993f-1d00-4dcb-bc60-3c0bff83546b req-fd3ddeb4-1374-493d-9385-e8cf86cd3ff9 service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] No waiting events found dispatching network-vif-plugged-382b237a-593e-47c2-99c0-18d3918d92a6 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1215.808175] env[62814]: WARNING nova.compute.manager [req-8302993f-1d00-4dcb-bc60-3c0bff83546b req-fd3ddeb4-1374-493d-9385-e8cf86cd3ff9 service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Received unexpected event network-vif-plugged-382b237a-593e-47c2-99c0-18d3918d92a6 for instance with vm_state building and task_state spawning. [ 1215.857889] env[62814]: DEBUG oslo_vmware.api [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294742, 'name': PowerOnVM_Task, 'duration_secs': 0.549786} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.858516] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1215.859040] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7487a5e2-34b2-49d2-b2af-0f71c1dfa2eb tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance 'b1d88997-e52f-41bd-b1b4-dd096d20d60a' progress to 100 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1215.868885] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31823c2a-dc4f-4c0f-91d5-a429010f5c38 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.892107] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91bf750b-74ec-4fd3-b816-4f01c8d9290d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.899812] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance '606e31eb-2349-427f-9c9b-ed9dc5b385f5' progress to 83 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1215.904599] env[62814]: DEBUG nova.network.neutron [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Successfully updated port: 382b237a-593e-47c2-99c0-18d3918d92a6 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1216.027804] env[62814]: DEBUG nova.compute.manager [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1216.029120] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1216.029120] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47b7bb30-ca00-4120-a3ed-730944a770ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.036754] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1216.036998] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d5ef913e-9eff-46af-9d8f-cf07dceefc26 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.043488] env[62814]: DEBUG oslo_vmware.api [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1216.043488] env[62814]: value = "task-4294743" [ 1216.043488] env[62814]: _type = "Task" [ 1216.043488] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.051011] env[62814]: DEBUG oslo_vmware.api [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.355669] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1216.355914] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1216.408058] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1216.409390] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1198275-6c9d-42c1-a9dd-7307fc178798 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.418601] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "refresh_cache-c9b67326-3058-4fa6-a094-b47bf8663444" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1216.418784] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "refresh_cache-c9b67326-3058-4fa6-a094-b47bf8663444" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1216.418972] env[62814]: DEBUG nova.network.neutron [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1216.422085] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1216.422085] env[62814]: value = "task-4294744" [ 1216.422085] env[62814]: _type = "Task" [ 1216.422085] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.424085] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b4d2cd-ff01-4fdd-a15d-cac3bfe57ebb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.437145] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6662d4-dba0-4464-8817-aa90f038e1b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.440333] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294744, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.469177] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a3b583-c2a8-40ad-a9cd-f199e7b87727 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.476053] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a943a5e-7ced-46a3-af0d-4591794e9365 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.489483] env[62814]: DEBUG nova.compute.provider_tree [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1216.553717] env[62814]: DEBUG oslo_vmware.api [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294743, 'name': PowerOffVM_Task, 'duration_secs': 0.218801} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.553717] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1216.554250] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1216.554250] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ade6bda-4b5b-42c2-b794-466551af52dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.616045] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1216.616356] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1216.616583] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Deleting the datastore file [datastore2] 33ab2d9e-e960-40ae-acf9-4fea0d9c830e {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1216.616849] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1ce0b4b-d9fa-45ae-b356-2d7a1cb7618f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1216.623881] env[62814]: DEBUG oslo_vmware.api [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for the task: (returnval){ [ 1216.623881] env[62814]: value = "task-4294746" [ 1216.623881] env[62814]: _type = "Task" [ 1216.623881] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1216.631769] env[62814]: DEBUG oslo_vmware.api [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294746, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.859076] env[62814]: DEBUG nova.compute.manager [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1216.932535] env[62814]: DEBUG oslo_vmware.api [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294744, 'name': PowerOnVM_Task, 'duration_secs': 0.448194} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.932852] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1216.933070] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bb918f4a-1b72-4252-840f-3667a356cc65 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance '606e31eb-2349-427f-9c9b-ed9dc5b385f5' progress to 100 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1216.966894] env[62814]: DEBUG nova.network.neutron [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1216.992229] env[62814]: DEBUG nova.scheduler.client.report [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1217.137417] env[62814]: DEBUG oslo_vmware.api [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Task: {'id': task-4294746, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146094} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1217.137707] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1217.137958] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1217.138330] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1217.138520] env[62814]: INFO nova.compute.manager [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1217.138779] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1217.138981] env[62814]: DEBUG nova.compute.manager [-] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1217.139206] env[62814]: DEBUG nova.network.neutron [-] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1217.162667] env[62814]: DEBUG nova.network.neutron [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Updating instance_info_cache with network_info: [{"id": "382b237a-593e-47c2-99c0-18d3918d92a6", "address": "fa:16:3e:9f:1f:03", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap382b237a-59", "ovs_interfaceid": "382b237a-593e-47c2-99c0-18d3918d92a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1217.383401] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1217.497462] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1217.497853] env[62814]: DEBUG nova.compute.manager [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1217.500451] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.199s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1217.500642] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1217.502758] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.799s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1217.504372] env[62814]: INFO nova.compute.claims [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1217.527459] env[62814]: INFO nova.scheduler.client.report [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted allocations for instance 40e3c00b-2129-476e-bc67-a0ef13781990 [ 1217.649197] env[62814]: DEBUG nova.compute.manager [req-d8a06eb9-825d-419d-94d2-59e2965314e7 req-99a8fd44-3645-4338-b391-0824a6b8e601 service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Received event network-vif-deleted-446b3cd1-e546-440b-bf99-c5bfb3dd4878 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1217.649479] env[62814]: INFO nova.compute.manager [req-d8a06eb9-825d-419d-94d2-59e2965314e7 req-99a8fd44-3645-4338-b391-0824a6b8e601 service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Neutron deleted interface 446b3cd1-e546-440b-bf99-c5bfb3dd4878; detaching it from the instance and deleting it from the info cache [ 1217.649717] env[62814]: DEBUG nova.network.neutron [req-d8a06eb9-825d-419d-94d2-59e2965314e7 req-99a8fd44-3645-4338-b391-0824a6b8e601 service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1217.664839] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "refresh_cache-c9b67326-3058-4fa6-a094-b47bf8663444" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1217.665303] env[62814]: DEBUG nova.compute.manager [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Instance network_info: |[{"id": "382b237a-593e-47c2-99c0-18d3918d92a6", "address": "fa:16:3e:9f:1f:03", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap382b237a-59", "ovs_interfaceid": "382b237a-593e-47c2-99c0-18d3918d92a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1217.665918] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9f:1f:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'afae6acd-1873-4228-9d5a-1cd5d4efe3e4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '382b237a-593e-47c2-99c0-18d3918d92a6', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1217.673464] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1217.674138] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1217.674384] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f063d780-3b02-46e2-9859-62546fd94303 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.693941] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1217.693941] env[62814]: value = "task-4294747" [ 1217.693941] env[62814]: _type = "Task" [ 1217.693941] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1217.703198] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294747, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1217.826023] env[62814]: DEBUG nova.network.neutron [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Port 0235e89b-ec0d-4439-817c-dd76d6987b80 binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1217.826023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.826215] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1217.826297] env[62814]: DEBUG nova.network.neutron [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1217.844895] env[62814]: DEBUG nova.compute.manager [req-43964b1a-e086-4a82-8d58-fed1324636e2 req-e2992024-afdb-40b4-8068-01f082162705 service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Received event network-changed-382b237a-593e-47c2-99c0-18d3918d92a6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1217.845185] env[62814]: DEBUG nova.compute.manager [req-43964b1a-e086-4a82-8d58-fed1324636e2 req-e2992024-afdb-40b4-8068-01f082162705 service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Refreshing instance network info cache due to event network-changed-382b237a-593e-47c2-99c0-18d3918d92a6. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1217.845485] env[62814]: DEBUG oslo_concurrency.lockutils [req-43964b1a-e086-4a82-8d58-fed1324636e2 req-e2992024-afdb-40b4-8068-01f082162705 service nova] Acquiring lock "refresh_cache-c9b67326-3058-4fa6-a094-b47bf8663444" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1217.845657] env[62814]: DEBUG oslo_concurrency.lockutils [req-43964b1a-e086-4a82-8d58-fed1324636e2 req-e2992024-afdb-40b4-8068-01f082162705 service nova] Acquired lock "refresh_cache-c9b67326-3058-4fa6-a094-b47bf8663444" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1217.845848] env[62814]: DEBUG nova.network.neutron [req-43964b1a-e086-4a82-8d58-fed1324636e2 req-e2992024-afdb-40b4-8068-01f082162705 service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Refreshing network info cache for port 382b237a-593e-47c2-99c0-18d3918d92a6 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1218.010736] env[62814]: DEBUG nova.compute.utils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1218.014855] env[62814]: DEBUG nova.compute.manager [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1218.015124] env[62814]: DEBUG nova.network.neutron [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1218.036308] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d769198b-2071-4839-99ad-97dc4ab9c88f tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "40e3c00b-2129-476e-bc67-a0ef13781990" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.277s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1218.057621] env[62814]: DEBUG nova.policy [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '64e84da705284d2dbf693a26ef184cd7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95c92336f9e746edba50b0b9e078b0dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1218.128030] env[62814]: DEBUG nova.network.neutron [-] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.152534] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21d280d9-19ed-40ae-9557-4eff78be6a2c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.164743] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed142153-5b2d-4b87-b790-399930df4c4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.201863] env[62814]: DEBUG nova.compute.manager [req-d8a06eb9-825d-419d-94d2-59e2965314e7 req-99a8fd44-3645-4338-b391-0824a6b8e601 service nova] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Detach interface failed, port_id=446b3cd1-e546-440b-bf99-c5bfb3dd4878, reason: Instance 33ab2d9e-e960-40ae-acf9-4fea0d9c830e could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1218.210618] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294747, 'name': CreateVM_Task, 'duration_secs': 0.371249} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.210788] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1218.211475] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.211645] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1218.211965] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1218.212242] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62c89595-5f63-4c75-bd53-219e2ddf865d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.216808] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1218.216808] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526a537c-0309-d9e3-b854-e118fb3fc389" [ 1218.216808] env[62814]: _type = "Task" [ 1218.216808] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.224801] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526a537c-0309-d9e3-b854-e118fb3fc389, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.359227] env[62814]: DEBUG nova.network.neutron [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Successfully created port: 0c147c05-6c14-4d99-ac3f-33be08394039 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1218.518818] env[62814]: DEBUG nova.compute.manager [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1218.631772] env[62814]: INFO nova.compute.manager [-] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Took 1.49 seconds to deallocate network for instance. [ 1218.728714] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]526a537c-0309-d9e3-b854-e118fb3fc389, 'name': SearchDatastore_Task, 'duration_secs': 0.01048} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.731105] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1218.731349] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1218.731587] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1218.731740] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1218.731973] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1218.732431] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c3612a1-707b-487f-866d-93b10d80c7d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.741575] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1218.741752] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1218.743295] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fddb9acd-ac5e-4d1d-bf00-b435770df639 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.747017] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6215d404-6ba1-4c8e-96d1-f7e3c4d078e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.752085] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1218.752085] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528705ec-eec5-68ac-c1f1-f96fd1fbd767" [ 1218.752085] env[62814]: _type = "Task" [ 1218.752085] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.757737] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f3b2a61-6230-400d-b3eb-eb0d8e30c0b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.761704] env[62814]: DEBUG nova.network.neutron [req-43964b1a-e086-4a82-8d58-fed1324636e2 req-e2992024-afdb-40b4-8068-01f082162705 service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Updated VIF entry in instance network info cache for port 382b237a-593e-47c2-99c0-18d3918d92a6. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1218.762154] env[62814]: DEBUG nova.network.neutron [req-43964b1a-e086-4a82-8d58-fed1324636e2 req-e2992024-afdb-40b4-8068-01f082162705 service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Updating instance_info_cache with network_info: [{"id": "382b237a-593e-47c2-99c0-18d3918d92a6", "address": "fa:16:3e:9f:1f:03", "network": {"id": "b2009dee-f767-469c-b460-85616eb66d4a", "bridge": "br-int", "label": "tempest-ServersTestJSON-1991175308-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d82993ef7dfa4d2f8f39db3577dd321a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "afae6acd-1873-4228-9d5a-1cd5d4efe3e4", "external-id": "nsx-vlan-transportzone-183", "segmentation_id": 183, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap382b237a-59", "ovs_interfaceid": "382b237a-593e-47c2-99c0-18d3918d92a6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.765318] env[62814]: DEBUG nova.network.neutron [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance_info_cache with network_info: [{"id": "0235e89b-ec0d-4439-817c-dd76d6987b80", "address": "fa:16:3e:46:9d:6d", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0235e89b-ec", "ovs_interfaceid": "0235e89b-ec0d-4439-817c-dd76d6987b80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1218.794658] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528705ec-eec5-68ac-c1f1-f96fd1fbd767, 'name': SearchDatastore_Task, 'duration_secs': 0.009365} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.797039] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255a1d55-b45c-4589-aed3-8302b6ff00ff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.799467] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14f88dc8-00b8-4d1f-995b-ecb71ece67a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.805878] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1218.805878] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52230ed7-b0e4-ce6f-505e-31ae91afd3a1" [ 1218.805878] env[62814]: _type = "Task" [ 1218.805878] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.811714] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a7d914-e406-49b0-950e-25e69dce96c7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.820603] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52230ed7-b0e4-ce6f-505e-31ae91afd3a1, 'name': SearchDatastore_Task, 'duration_secs': 0.009775} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.828765] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1218.829070] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c9b67326-3058-4fa6-a094-b47bf8663444/c9b67326-3058-4fa6-a094-b47bf8663444.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1218.830215] env[62814]: DEBUG nova.compute.provider_tree [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1218.831171] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-92255556-e82b-4ab7-8960-f7e2cf4d584c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.842020] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1218.842020] env[62814]: value = "task-4294748" [ 1218.842020] env[62814]: _type = "Task" [ 1218.842020] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.847838] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294748, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.139375] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1219.155851] env[62814]: DEBUG nova.network.neutron [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Port 837d03b5-250f-44d2-bafb-31184267117d binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1219.156144] env[62814]: DEBUG oslo_concurrency.lockutils [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1219.156302] env[62814]: DEBUG oslo_concurrency.lockutils [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1219.156550] env[62814]: DEBUG nova.network.neutron [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1219.264803] env[62814]: DEBUG oslo_concurrency.lockutils [req-43964b1a-e086-4a82-8d58-fed1324636e2 req-e2992024-afdb-40b4-8068-01f082162705 service nova] Releasing lock "refresh_cache-c9b67326-3058-4fa6-a094-b47bf8663444" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1219.296896] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1219.334434] env[62814]: DEBUG nova.scheduler.client.report [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1219.351040] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294748, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.529089] env[62814]: DEBUG nova.compute.manager [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1219.561044] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1219.562029] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1219.562029] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1219.562029] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1219.562029] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1219.562029] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1219.562301] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1219.562555] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1219.562775] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1219.563011] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1219.563226] env[62814]: DEBUG nova.virt.hardware [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1219.564159] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bae622a-5956-4d83-9705-7a344a55e5ee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.572760] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cfa047-4db1-4a18-b63c-aa1a89e78186 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.800968] env[62814]: DEBUG nova.compute.manager [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62814) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:901}} [ 1219.839939] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1219.842053] env[62814]: DEBUG nova.compute.manager [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1219.846325] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.463s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1219.847862] env[62814]: INFO nova.compute.claims [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1219.863908] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294748, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547891} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.864208] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] c9b67326-3058-4fa6-a094-b47bf8663444/c9b67326-3058-4fa6-a094-b47bf8663444.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1219.864432] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1219.864688] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-637b098d-14e3-4001-97c4-93c812fc9d40 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1219.871284] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1219.871284] env[62814]: value = "task-4294749" [ 1219.871284] env[62814]: _type = "Task" [ 1219.871284] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1219.879238] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294749, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.944204] env[62814]: DEBUG nova.compute.manager [req-4c7d97fb-f8ff-4bb9-86c1-9185dfe5675b req-b8b93f16-1cb6-43f2-8f3a-e1921ecef0bb service nova] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Received event network-vif-plugged-0c147c05-6c14-4d99-ac3f-33be08394039 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1219.944443] env[62814]: DEBUG oslo_concurrency.lockutils [req-4c7d97fb-f8ff-4bb9-86c1-9185dfe5675b req-b8b93f16-1cb6-43f2-8f3a-e1921ecef0bb service nova] Acquiring lock "fed47b28-64fb-4af9-9f8e-97a63afda514-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1219.944650] env[62814]: DEBUG oslo_concurrency.lockutils [req-4c7d97fb-f8ff-4bb9-86c1-9185dfe5675b req-b8b93f16-1cb6-43f2-8f3a-e1921ecef0bb service nova] Lock "fed47b28-64fb-4af9-9f8e-97a63afda514-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1219.944821] env[62814]: DEBUG oslo_concurrency.lockutils [req-4c7d97fb-f8ff-4bb9-86c1-9185dfe5675b req-b8b93f16-1cb6-43f2-8f3a-e1921ecef0bb service nova] Lock "fed47b28-64fb-4af9-9f8e-97a63afda514-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1219.944986] env[62814]: DEBUG nova.compute.manager [req-4c7d97fb-f8ff-4bb9-86c1-9185dfe5675b req-b8b93f16-1cb6-43f2-8f3a-e1921ecef0bb service nova] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] No waiting events found dispatching network-vif-plugged-0c147c05-6c14-4d99-ac3f-33be08394039 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1219.945483] env[62814]: WARNING nova.compute.manager [req-4c7d97fb-f8ff-4bb9-86c1-9185dfe5675b req-b8b93f16-1cb6-43f2-8f3a-e1921ecef0bb service nova] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Received unexpected event network-vif-plugged-0c147c05-6c14-4d99-ac3f-33be08394039 for instance with vm_state building and task_state spawning. [ 1219.958282] env[62814]: DEBUG nova.network.neutron [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance_info_cache with network_info: [{"id": "837d03b5-250f-44d2-bafb-31184267117d", "address": "fa:16:3e:24:ee:bf", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap837d03b5-25", "ovs_interfaceid": "837d03b5-250f-44d2-bafb-31184267117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1220.041940] env[62814]: DEBUG nova.network.neutron [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Successfully updated port: 0c147c05-6c14-4d99-ac3f-33be08394039 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1220.353286] env[62814]: DEBUG nova.compute.utils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1220.354735] env[62814]: DEBUG nova.compute.manager [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1220.354910] env[62814]: DEBUG nova.network.neutron [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1220.380607] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294749, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.296493} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.380863] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1220.381616] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23e5daf-24b8-49a8-bf30-fa4fed743c73 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.405420] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] c9b67326-3058-4fa6-a094-b47bf8663444/c9b67326-3058-4fa6-a094-b47bf8663444.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1220.405955] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b511b57-7953-4fcb-bfd4-1c19d0ea9f4f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.420904] env[62814]: DEBUG nova.policy [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5779eaa60ec44b2b80660e521b598eb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10018060297c409d82935ad626cc54c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1220.427677] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1220.427677] env[62814]: value = "task-4294750" [ 1220.427677] env[62814]: _type = "Task" [ 1220.427677] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.437404] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294750, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.461361] env[62814]: DEBUG oslo_concurrency.lockutils [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1220.545509] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "refresh_cache-fed47b28-64fb-4af9-9f8e-97a63afda514" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.545509] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "refresh_cache-fed47b28-64fb-4af9-9f8e-97a63afda514" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1220.545509] env[62814]: DEBUG nova.network.neutron [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1220.759418] env[62814]: DEBUG nova.network.neutron [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Successfully created port: 9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1220.858296] env[62814]: DEBUG nova.compute.manager [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1220.934967] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1220.939225] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.964995] env[62814]: DEBUG nova.compute.manager [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62814) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:901}} [ 1220.965554] env[62814]: DEBUG oslo_concurrency.lockutils [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1221.040524] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420dfe7d-12ba-42bd-9e30-d615372e60b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.050510] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9983ee9-e6a1-45a0-bfef-d581ad78ac7a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.081460] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa7077c-9c7f-4607-b44b-404e566ce631 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.085229] env[62814]: DEBUG nova.network.neutron [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1221.090555] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960526c5-dd10-454b-9b1a-283daedac5e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.104993] env[62814]: DEBUG nova.compute.provider_tree [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1221.226918] env[62814]: DEBUG nova.network.neutron [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Updating instance_info_cache with network_info: [{"id": "0c147c05-6c14-4d99-ac3f-33be08394039", "address": "fa:16:3e:11:52:08", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c147c05-6c", "ovs_interfaceid": "0c147c05-6c14-4d99-ac3f-33be08394039", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1221.439334] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294750, 'name': ReconfigVM_Task, 'duration_secs': 0.766593} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.439657] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Reconfigured VM instance instance-00000072 to attach disk [datastore2] c9b67326-3058-4fa6-a094-b47bf8663444/c9b67326-3058-4fa6-a094-b47bf8663444.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1221.440351] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39924159-f544-441a-a941-a5b114781e90 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.447074] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1221.447074] env[62814]: value = "task-4294751" [ 1221.447074] env[62814]: _type = "Task" [ 1221.447074] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.455814] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294751, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.608104] env[62814]: DEBUG nova.scheduler.client.report [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1221.729340] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "refresh_cache-fed47b28-64fb-4af9-9f8e-97a63afda514" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1221.729814] env[62814]: DEBUG nova.compute.manager [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Instance network_info: |[{"id": "0c147c05-6c14-4d99-ac3f-33be08394039", "address": "fa:16:3e:11:52:08", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c147c05-6c", "ovs_interfaceid": "0c147c05-6c14-4d99-ac3f-33be08394039", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1221.730283] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:11:52:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '605f83bd-808c-4b54-922e-54b14690987a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c147c05-6c14-4d99-ac3f-33be08394039', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1221.738693] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1221.739736] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1221.739736] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca6387e4-9ca5-4998-a98a-6d0e79fe0e5b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.759141] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1221.759141] env[62814]: value = "task-4294752" [ 1221.759141] env[62814]: _type = "Task" [ 1221.759141] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.769374] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294752, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.868764] env[62814]: DEBUG nova.compute.manager [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1221.902461] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1221.903765] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1221.904015] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1221.904283] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1221.904452] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1221.904595] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1221.905106] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1221.905383] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1221.905500] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1221.905709] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1221.906090] env[62814]: DEBUG nova.virt.hardware [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1221.907662] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1fb6e1-4be0-4523-bbb3-b53e7d0878a9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.916547] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200c8288-adbd-4592-a918-0fee1ed9008b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.956721] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294751, 'name': Rename_Task, 'duration_secs': 0.211889} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.957042] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1221.957322] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d596a6dc-4d55-47b9-98ef-5251d5a0c451 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.963756] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1221.963756] env[62814]: value = "task-4294753" [ 1221.963756] env[62814]: _type = "Task" [ 1221.963756] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.971651] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294753, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.973798] env[62814]: DEBUG nova.compute.manager [req-41635f95-421a-492e-a28c-908db201ed9b req-df294a80-c9fe-4db4-8796-d2b03a3a264d service nova] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Received event network-changed-0c147c05-6c14-4d99-ac3f-33be08394039 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1221.974012] env[62814]: DEBUG nova.compute.manager [req-41635f95-421a-492e-a28c-908db201ed9b req-df294a80-c9fe-4db4-8796-d2b03a3a264d service nova] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Refreshing instance network info cache due to event network-changed-0c147c05-6c14-4d99-ac3f-33be08394039. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1221.974265] env[62814]: DEBUG oslo_concurrency.lockutils [req-41635f95-421a-492e-a28c-908db201ed9b req-df294a80-c9fe-4db4-8796-d2b03a3a264d service nova] Acquiring lock "refresh_cache-fed47b28-64fb-4af9-9f8e-97a63afda514" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.974405] env[62814]: DEBUG oslo_concurrency.lockutils [req-41635f95-421a-492e-a28c-908db201ed9b req-df294a80-c9fe-4db4-8796-d2b03a3a264d service nova] Acquired lock "refresh_cache-fed47b28-64fb-4af9-9f8e-97a63afda514" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1221.974568] env[62814]: DEBUG nova.network.neutron [req-41635f95-421a-492e-a28c-908db201ed9b req-df294a80-c9fe-4db4-8796-d2b03a3a264d service nova] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Refreshing network info cache for port 0c147c05-6c14-4d99-ac3f-33be08394039 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1222.113423] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1222.113966] env[62814]: DEBUG nova.compute.manager [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1222.117364] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.978s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1222.117490] env[62814]: DEBUG nova.objects.instance [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lazy-loading 'resources' on Instance uuid 33ab2d9e-e960-40ae-acf9-4fea0d9c830e {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1222.271591] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294752, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.473749] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.508121] env[62814]: DEBUG nova.network.neutron [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Successfully updated port: 9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1222.620343] env[62814]: DEBUG nova.compute.utils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1222.626260] env[62814]: DEBUG nova.compute.manager [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1222.626611] env[62814]: DEBUG nova.network.neutron [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1222.666015] env[62814]: DEBUG nova.policy [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e246f49b0b84cd093549b6d6b384e68', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6e52480dd2c467790622901940cf385', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1222.678935] env[62814]: DEBUG nova.network.neutron [req-41635f95-421a-492e-a28c-908db201ed9b req-df294a80-c9fe-4db4-8796-d2b03a3a264d service nova] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Updated VIF entry in instance network info cache for port 0c147c05-6c14-4d99-ac3f-33be08394039. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1222.679173] env[62814]: DEBUG nova.network.neutron [req-41635f95-421a-492e-a28c-908db201ed9b req-df294a80-c9fe-4db4-8796-d2b03a3a264d service nova] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Updating instance_info_cache with network_info: [{"id": "0c147c05-6c14-4d99-ac3f-33be08394039", "address": "fa:16:3e:11:52:08", "network": {"id": "7243efaf-643e-4f7d-aab0-9ef43ae4833d", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1356533893-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "95c92336f9e746edba50b0b9e078b0dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "605f83bd-808c-4b54-922e-54b14690987a", "external-id": "nsx-vlan-transportzone-25", "segmentation_id": 25, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c147c05-6c", "ovs_interfaceid": "0c147c05-6c14-4d99-ac3f-33be08394039", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1222.773289] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294752, 'name': CreateVM_Task, 'duration_secs': 0.61745} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.773474] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1222.782502] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1222.782682] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1222.783013] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1222.783460] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50435c98-58ec-4c7b-be31-28ec0c775bfb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.788572] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1222.788572] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5229db4d-1f50-28d2-5c64-143751b45632" [ 1222.788572] env[62814]: _type = "Task" [ 1222.788572] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.797987] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5229db4d-1f50-28d2-5c64-143751b45632, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.808055] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0993b0-3c3d-4ac9-a722-ef44496fdbad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.815640] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69f1fe8b-0d9e-48f6-ae07-616731ef00b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.847176] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3714a311-52a9-4147-80e3-487c2e66991a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.854823] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66604c9f-f093-4a10-9b59-e9acf3d8c4cf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.868763] env[62814]: DEBUG nova.compute.provider_tree [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1222.974765] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294753, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.014431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1223.014431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1223.014431] env[62814]: DEBUG nova.network.neutron [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1223.051910] env[62814]: DEBUG nova.network.neutron [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Successfully created port: 3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1223.127811] env[62814]: DEBUG nova.compute.manager [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1223.186829] env[62814]: DEBUG oslo_concurrency.lockutils [req-41635f95-421a-492e-a28c-908db201ed9b req-df294a80-c9fe-4db4-8796-d2b03a3a264d service nova] Releasing lock "refresh_cache-fed47b28-64fb-4af9-9f8e-97a63afda514" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1223.301137] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5229db4d-1f50-28d2-5c64-143751b45632, 'name': SearchDatastore_Task, 'duration_secs': 0.011222} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.301275] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1223.301437] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1223.301671] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1223.301812] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1223.301986] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1223.302276] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b07b0c7b-21de-4f61-a054-84f98508d167 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.312617] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1223.312708] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1223.313447] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efd1139f-b922-45ef-9797-38c3769c279c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.321837] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1223.321837] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5234f823-9db0-2f6f-678e-02ff83046546" [ 1223.321837] env[62814]: _type = "Task" [ 1223.321837] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.329763] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5234f823-9db0-2f6f-678e-02ff83046546, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.374022] env[62814]: DEBUG nova.scheduler.client.report [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1223.476056] env[62814]: DEBUG oslo_vmware.api [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294753, 'name': PowerOnVM_Task, 'duration_secs': 1.141084} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.476056] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1223.476056] env[62814]: INFO nova.compute.manager [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Took 7.98 seconds to spawn the instance on the hypervisor. [ 1223.476315] env[62814]: DEBUG nova.compute.manager [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1223.476966] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8a3eda-978f-4aa6-92b0-c87c3eafa594 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.547125] env[62814]: DEBUG nova.network.neutron [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1223.694612] env[62814]: DEBUG nova.network.neutron [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [{"id": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "address": "fa:16:3e:ff:ce:8d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9edabfa8-97", "ovs_interfaceid": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1223.833094] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5234f823-9db0-2f6f-678e-02ff83046546, 'name': SearchDatastore_Task, 'duration_secs': 0.009189} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.834242] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42bbc94d-241e-47dd-8703-9387e5c682fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.839822] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1223.839822] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e40adf-595d-2533-9b07-c159aa8d1075" [ 1223.839822] env[62814]: _type = "Task" [ 1223.839822] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.847442] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e40adf-595d-2533-9b07-c159aa8d1075, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.877349] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1223.879635] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 2.945s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1223.901955] env[62814]: INFO nova.scheduler.client.report [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Deleted allocations for instance 33ab2d9e-e960-40ae-acf9-4fea0d9c830e [ 1223.993553] env[62814]: INFO nova.compute.manager [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Took 17.37 seconds to build instance. [ 1223.996549] env[62814]: DEBUG nova.compute.manager [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received event network-vif-plugged-9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1223.996761] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] Acquiring lock "b5a2fb6c-7078-4f3c-b511-014beb96391d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1223.996970] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] Lock "b5a2fb6c-7078-4f3c-b511-014beb96391d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1223.997253] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] Lock "b5a2fb6c-7078-4f3c-b511-014beb96391d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1223.997529] env[62814]: DEBUG nova.compute.manager [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] No waiting events found dispatching network-vif-plugged-9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1223.997727] env[62814]: WARNING nova.compute.manager [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received unexpected event network-vif-plugged-9edabfa8-9760-4ef4-adfb-afc6e2200b50 for instance with vm_state building and task_state spawning. [ 1223.997907] env[62814]: DEBUG nova.compute.manager [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received event network-changed-9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1223.998099] env[62814]: DEBUG nova.compute.manager [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing instance network info cache due to event network-changed-9edabfa8-9760-4ef4-adfb-afc6e2200b50. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1223.998324] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] Acquiring lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.049516] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.139531] env[62814]: DEBUG nova.compute.manager [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1224.167080] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1224.167411] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1224.168229] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1224.168229] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1224.168229] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1224.168386] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1224.168497] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1224.168629] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1224.168795] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1224.168993] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1224.169228] env[62814]: DEBUG nova.virt.hardware [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1224.170122] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b23b358d-659a-441a-93c9-9361d954355e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.178617] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc4fc8ff-eb16-471b-813a-b498014f3ce7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.196897] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1224.197219] env[62814]: DEBUG nova.compute.manager [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Instance network_info: |[{"id": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "address": "fa:16:3e:ff:ce:8d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9edabfa8-97", "ovs_interfaceid": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1224.197505] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] Acquired lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.197679] env[62814]: DEBUG nova.network.neutron [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing network info cache for port 9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1224.198927] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:ce:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe99da4f-5630-4afd-918b-b327193d8489', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9edabfa8-9760-4ef4-adfb-afc6e2200b50', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1224.206318] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1224.209071] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1224.209553] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04689acb-1467-461d-a5ad-b78d5c0d60da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.229838] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1224.229838] env[62814]: value = "task-4294754" [ 1224.229838] env[62814]: _type = "Task" [ 1224.229838] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.238329] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294754, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.349851] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e40adf-595d-2533-9b07-c159aa8d1075, 'name': SearchDatastore_Task, 'duration_secs': 0.011584} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.351150] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1224.351150] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] fed47b28-64fb-4af9-9f8e-97a63afda514/fed47b28-64fb-4af9-9f8e-97a63afda514.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1224.351150] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9e976c7-8634-4edb-b223-e92d927bebf2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.357520] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1224.357520] env[62814]: value = "task-4294755" [ 1224.357520] env[62814]: _type = "Task" [ 1224.357520] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.365102] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294755, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.383492] env[62814]: DEBUG nova.objects.instance [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'migration_context' on Instance uuid b1d88997-e52f-41bd-b1b4-dd096d20d60a {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1224.409567] env[62814]: DEBUG oslo_concurrency.lockutils [None req-00465437-b93b-4ae6-9810-ac7efbbec186 tempest-AttachVolumeTestJSON-1694269603 tempest-AttachVolumeTestJSON-1694269603-project-member] Lock "33ab2d9e-e960-40ae-acf9-4fea0d9c830e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.900s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.476961] env[62814]: DEBUG nova.network.neutron [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updated VIF entry in instance network info cache for port 9edabfa8-9760-4ef4-adfb-afc6e2200b50. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1224.477475] env[62814]: DEBUG nova.network.neutron [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [{"id": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "address": "fa:16:3e:ff:ce:8d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9edabfa8-97", "ovs_interfaceid": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1224.500214] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d3de8a1a-0201-422f-8baa-333d1dcc319d tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "c9b67326-3058-4fa6-a094-b47bf8663444" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.902s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1224.740876] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294754, 'name': CreateVM_Task, 'duration_secs': 0.333406} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.741866] env[62814]: DEBUG nova.network.neutron [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Successfully updated port: 3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1224.743081] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1224.744807] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1224.746031] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1224.746031] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1224.746031] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb53bd24-20e2-4452-8d58-39c26963400f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.751107] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1224.751107] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521f9388-09d1-ffc7-3af6-9fbfd466bbe2" [ 1224.751107] env[62814]: _type = "Task" [ 1224.751107] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.762133] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521f9388-09d1-ffc7-3af6-9fbfd466bbe2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.867832] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294755, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.982258] env[62814]: DEBUG oslo_concurrency.lockutils [req-8a599972-d697-4349-b2b7-7dd94cc33709 req-387af9c4-ea93-412a-83a8-63b3c65f1812 service nova] Releasing lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1225.030908] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef579579-b558-48f2-8753-15020d185b9f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.038346] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6ab957-029d-4298-999a-b14e6a80df82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.069787] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730380f1-7b92-424d-a94a-ac1d68788feb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.076965] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f083c56d-26a6-40c9-9e72-9638999393fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.089907] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "c9b67326-3058-4fa6-a094-b47bf8663444" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1225.090200] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "c9b67326-3058-4fa6-a094-b47bf8663444" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1225.090388] env[62814]: DEBUG nova.compute.manager [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1225.090831] env[62814]: DEBUG nova.compute.provider_tree [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1225.092583] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db15380-2533-4014-b59b-c9b33364db5a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.098548] env[62814]: DEBUG nova.compute.manager [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1225.099116] env[62814]: DEBUG nova.objects.instance [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lazy-loading 'flavor' on Instance uuid c9b67326-3058-4fa6-a094-b47bf8663444 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1225.247544] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.247702] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1225.247907] env[62814]: DEBUG nova.network.neutron [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1225.262297] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521f9388-09d1-ffc7-3af6-9fbfd466bbe2, 'name': SearchDatastore_Task, 'duration_secs': 0.056365} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.263175] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1225.263411] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1225.263638] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1225.263785] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1225.263964] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1225.264507] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d573619e-d7f1-40e2-bc5a-7e65ff1bb469 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.273777] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1225.273971] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1225.274784] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-849de5d5-5d0b-4454-b00d-9ed1172b0723 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.280374] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1225.280374] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52af8601-73ab-cf93-ee65-5db49e286e51" [ 1225.280374] env[62814]: _type = "Task" [ 1225.280374] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.287893] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52af8601-73ab-cf93-ee65-5db49e286e51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.369086] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294755, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.538261} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.369354] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] fed47b28-64fb-4af9-9f8e-97a63afda514/fed47b28-64fb-4af9-9f8e-97a63afda514.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1225.369574] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1225.369823] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b88c396e-561f-4855-b578-045a7652d538 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.376280] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1225.376280] env[62814]: value = "task-4294757" [ 1225.376280] env[62814]: _type = "Task" [ 1225.376280] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.385502] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294757, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.502496] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1225.506434] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1225.506584] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 1225.596798] env[62814]: DEBUG nova.scheduler.client.report [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1225.790868] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52af8601-73ab-cf93-ee65-5db49e286e51, 'name': SearchDatastore_Task, 'duration_secs': 0.009535} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.791712] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-943325d3-d8e8-47de-8747-d7f033466237 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.795590] env[62814]: DEBUG nova.network.neutron [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1225.798990] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1225.798990] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52952c09-74b2-ff70-00e2-82e8c497b51b" [ 1225.798990] env[62814]: _type = "Task" [ 1225.798990] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.807199] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52952c09-74b2-ff70-00e2-82e8c497b51b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.891585] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294757, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07598} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.891858] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1225.892687] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969a8f73-6069-49a1-8409-76de21b2c8c7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.917202] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Reconfiguring VM instance instance-00000073 to attach disk [datastore2] fed47b28-64fb-4af9-9f8e-97a63afda514/fed47b28-64fb-4af9-9f8e-97a63afda514.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1225.917570] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97fe8ea0-f07a-4426-a802-ea0dd3453739 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.936522] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1225.936522] env[62814]: value = "task-4294758" [ 1225.936522] env[62814]: _type = "Task" [ 1225.936522] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1225.945982] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294758, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.975298] env[62814]: DEBUG nova.network.neutron [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance_info_cache with network_info: [{"id": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "address": "fa:16:3e:94:da:3d", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d17bc61-a2", "ovs_interfaceid": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1226.028438] env[62814]: DEBUG nova.compute.manager [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Received event network-vif-plugged-3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1226.028438] env[62814]: DEBUG oslo_concurrency.lockutils [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] Acquiring lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1226.028783] env[62814]: DEBUG oslo_concurrency.lockutils [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1226.028783] env[62814]: DEBUG oslo_concurrency.lockutils [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.028860] env[62814]: DEBUG nova.compute.manager [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] No waiting events found dispatching network-vif-plugged-3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1226.028991] env[62814]: WARNING nova.compute.manager [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Received unexpected event network-vif-plugged-3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8 for instance with vm_state building and task_state spawning. [ 1226.029174] env[62814]: DEBUG nova.compute.manager [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Received event network-changed-3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1226.029331] env[62814]: DEBUG nova.compute.manager [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Refreshing instance network info cache due to event network-changed-3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1226.029495] env[62814]: DEBUG oslo_concurrency.lockutils [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] Acquiring lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1226.106951] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1226.107426] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5416a23-e79b-49d3-bdcd-942a33e74d30 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.115402] env[62814]: DEBUG oslo_vmware.api [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1226.115402] env[62814]: value = "task-4294759" [ 1226.115402] env[62814]: _type = "Task" [ 1226.115402] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.124101] env[62814]: DEBUG oslo_vmware.api [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294759, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.309155] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52952c09-74b2-ff70-00e2-82e8c497b51b, 'name': SearchDatastore_Task, 'duration_secs': 0.010027} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.309433] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1226.309692] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b5a2fb6c-7078-4f3c-b511-014beb96391d/b5a2fb6c-7078-4f3c-b511-014beb96391d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1226.309953] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c689d4c-5070-4532-9eed-1e5fb8d33a26 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.317914] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1226.317914] env[62814]: value = "task-4294760" [ 1226.317914] env[62814]: _type = "Task" [ 1226.317914] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.327030] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294760, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.447052] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294758, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.478035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1226.478220] env[62814]: DEBUG nova.compute.manager [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Instance network_info: |[{"id": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "address": "fa:16:3e:94:da:3d", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d17bc61-a2", "ovs_interfaceid": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1226.478572] env[62814]: DEBUG oslo_concurrency.lockutils [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] Acquired lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1226.478802] env[62814]: DEBUG nova.network.neutron [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Refreshing network info cache for port 3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1226.480187] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:94:da:3d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '51bac3c3-00ab-4a07-9e28-b3c951dee565', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1226.489707] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1226.490354] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1226.491381] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c000ce91-3353-488c-bee5-ef577695b295 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.509741] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1226.509922] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1226.517607] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1226.517607] env[62814]: value = "task-4294761" [ 1226.517607] env[62814]: _type = "Task" [ 1226.517607] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.526461] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294761, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.610306] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.730s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1226.615913] env[62814]: DEBUG oslo_concurrency.lockutils [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 5.651s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1226.633646] env[62814]: DEBUG oslo_vmware.api [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294759, 'name': PowerOffVM_Task, 'duration_secs': 0.30106} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.634613] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1226.634810] env[62814]: DEBUG nova.compute.manager [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1226.635750] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a29fa05-c762-4a15-9d8e-ed94ba3566aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.829396] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294760, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.946982] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294758, 'name': ReconfigVM_Task, 'duration_secs': 0.56397} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.947355] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Reconfigured VM instance instance-00000073 to attach disk [datastore2] fed47b28-64fb-4af9-9f8e-97a63afda514/fed47b28-64fb-4af9-9f8e-97a63afda514.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1226.948069] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2955a3c1-516b-4bcd-acb7-eb4a62c766ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.954615] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1226.954615] env[62814]: value = "task-4294762" [ 1226.954615] env[62814]: _type = "Task" [ 1226.954615] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.962778] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294762, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.013242] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1227.030418] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294761, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.127475] env[62814]: DEBUG nova.objects.instance [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'migration_context' on Instance uuid 606e31eb-2349-427f-9c9b-ed9dc5b385f5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1227.148884] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8a1fc211-fe67-490d-8db4-41e28f54fb8f tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "c9b67326-3058-4fa6-a094-b47bf8663444" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.059s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1227.183719] env[62814]: DEBUG nova.network.neutron [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updated VIF entry in instance network info cache for port 3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1227.184115] env[62814]: DEBUG nova.network.neutron [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance_info_cache with network_info: [{"id": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "address": "fa:16:3e:94:da:3d", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d17bc61-a2", "ovs_interfaceid": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1227.330557] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294760, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56959} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.330911] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] b5a2fb6c-7078-4f3c-b511-014beb96391d/b5a2fb6c-7078-4f3c-b511-014beb96391d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1227.331176] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1227.331880] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9afef18e-3dc2-47a5-8e18-2334097f55f8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.339485] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1227.339485] env[62814]: value = "task-4294763" [ 1227.339485] env[62814]: _type = "Task" [ 1227.339485] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.350155] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294763, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.464606] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294762, 'name': Rename_Task, 'duration_secs': 0.170543} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.464911] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1227.465230] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d860ce18-b1d5-403e-8dde-8a4417b54556 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.471943] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1227.471943] env[62814]: value = "task-4294764" [ 1227.471943] env[62814]: _type = "Task" [ 1227.471943] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.479773] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294764, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.527904] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294761, 'name': CreateVM_Task, 'duration_secs': 0.523845} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.528111] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1227.528788] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1227.528960] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1227.529315] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1227.529574] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4a2a79a-66dd-4b48-b357-c252a0a75e70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.533977] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1227.533977] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520006ec-c66a-21b3-ac4f-40160dbc3400" [ 1227.533977] env[62814]: _type = "Task" [ 1227.533977] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.541338] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520006ec-c66a-21b3-ac4f-40160dbc3400, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.686745] env[62814]: DEBUG oslo_concurrency.lockutils [req-14d8f229-dce4-4631-bd38-08374e4ac76c req-bd9dc51f-befc-4bf3-aefb-a8fb94c16e51 service nova] Releasing lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1227.780439] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6c78db-a033-44dc-a267-9eebf76397ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.788681] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-991a2ebf-38b8-4b3e-9c6c-7a47a07676e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.818746] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5769712c-8537-451f-bc03-44b8d4639291 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.825250] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "c9b67326-3058-4fa6-a094-b47bf8663444" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1227.825493] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "c9b67326-3058-4fa6-a094-b47bf8663444" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1227.825696] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "c9b67326-3058-4fa6-a094-b47bf8663444-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1227.825874] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "c9b67326-3058-4fa6-a094-b47bf8663444-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1227.826052] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "c9b67326-3058-4fa6-a094-b47bf8663444-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1227.828606] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d4aa5cc-199b-4873-9e4a-db0371cab29d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.832815] env[62814]: INFO nova.compute.manager [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Terminating instance [ 1227.843990] env[62814]: DEBUG nova.compute.provider_tree [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1227.855824] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294763, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062648} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.855999] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1227.856732] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e6fc827-bcee-44b9-8af8-2ab8c491c33f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.878245] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Reconfiguring VM instance instance-00000074 to attach disk [datastore2] b5a2fb6c-7078-4f3c-b511-014beb96391d/b5a2fb6c-7078-4f3c-b511-014beb96391d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1227.879164] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23dd0a68-a301-426d-861c-e0c9b77fe82c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.898708] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1227.898708] env[62814]: value = "task-4294765" [ 1227.898708] env[62814]: _type = "Task" [ 1227.898708] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.906787] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294765, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.981918] env[62814]: DEBUG oslo_vmware.api [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294764, 'name': PowerOnVM_Task, 'duration_secs': 0.470014} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.982229] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1227.982445] env[62814]: INFO nova.compute.manager [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Took 8.45 seconds to spawn the instance on the hypervisor. [ 1227.982625] env[62814]: DEBUG nova.compute.manager [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1227.983400] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-268bf47e-794e-425f-b929-0de69d848b40 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.044162] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]520006ec-c66a-21b3-ac4f-40160dbc3400, 'name': SearchDatastore_Task, 'duration_secs': 0.009593} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.044525] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1228.044652] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1228.044886] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.045046] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1228.045231] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1228.045520] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-004c4c7a-0b25-4f03-b240-ae2bd10cac5e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.053048] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1228.053221] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1228.053887] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-70b4c33a-fe3e-421b-8047-8008095ef696 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.058659] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1228.058659] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529541ad-e8a0-7f33-053e-996e5493b0b9" [ 1228.058659] env[62814]: _type = "Task" [ 1228.058659] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.065901] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529541ad-e8a0-7f33-053e-996e5493b0b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.153188] env[62814]: INFO nova.compute.manager [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Swapping old allocation on dict_keys(['7136a6f7-3927-4aa1-a4e9-7fcbd9976745']) held by migration 7c79fbaa-6c2c-4243-8ae4-fa861c910052 for instance [ 1228.175561] env[62814]: DEBUG nova.scheduler.client.report [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Overwriting current allocation {'allocations': {'7136a6f7-3927-4aa1-a4e9-7fcbd9976745': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 179}}, 'project_id': '446e63a428c34d38ad07d01f0c90b23c', 'user_id': '9df1642a53c940b2bfacc82e1ac975d0', 'consumer_generation': 1} on consumer b1d88997-e52f-41bd-b1b4-dd096d20d60a {{(pid=62814) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1228.267419] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1228.267620] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1228.267804] env[62814]: DEBUG nova.network.neutron [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1228.337097] env[62814]: DEBUG nova.compute.manager [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1228.337256] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1228.338110] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c593d920-7164-4f84-977a-9a9a31cad2eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.345952] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1228.346196] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b6bf905-c692-4aa1-a34e-0359f7b1b4d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.366647] env[62814]: ERROR nova.scheduler.client.report [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [req-1ce1facd-ba2d-4477-9a34-74a45b382b8c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1ce1facd-ba2d-4477-9a34-74a45b382b8c"}]} [ 1228.384459] env[62814]: DEBUG nova.scheduler.client.report [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1228.397959] env[62814]: DEBUG nova.scheduler.client.report [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1228.398207] env[62814]: DEBUG nova.compute.provider_tree [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1228.408046] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1228.408448] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1228.408523] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleting the datastore file [datastore2] c9b67326-3058-4fa6-a094-b47bf8663444 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1228.412079] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7dcc14a-e62a-48df-ae64-8535200a1fb5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.414445] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294765, 'name': ReconfigVM_Task, 'duration_secs': 0.335474} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.415393] env[62814]: DEBUG nova.scheduler.client.report [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1228.417336] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Reconfigured VM instance instance-00000074 to attach disk [datastore2] b5a2fb6c-7078-4f3c-b511-014beb96391d/b5a2fb6c-7078-4f3c-b511-014beb96391d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1228.418490] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d8d772b-5401-4544-b89c-f735b7699a64 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.422766] env[62814]: DEBUG oslo_vmware.api [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1228.422766] env[62814]: value = "task-4294767" [ 1228.422766] env[62814]: _type = "Task" [ 1228.422766] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.427426] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1228.427426] env[62814]: value = "task-4294768" [ 1228.427426] env[62814]: _type = "Task" [ 1228.427426] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.433316] env[62814]: DEBUG oslo_vmware.api [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294767, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.438591] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294768, 'name': Rename_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.439131] env[62814]: DEBUG nova.scheduler.client.report [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1228.502682] env[62814]: INFO nova.compute.manager [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Took 18.01 seconds to build instance. [ 1228.572405] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529541ad-e8a0-7f33-053e-996e5493b0b9, 'name': SearchDatastore_Task, 'duration_secs': 0.010696} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.576044] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91805fb3-40a3-4395-ac87-7cfef99f6fe1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.580971] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1228.580971] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dc8fda-63d8-158a-b11c-9de2d8db3bff" [ 1228.580971] env[62814]: _type = "Task" [ 1228.580971] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.586064] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a0f14d3-d7a7-4873-a581-e388333a3fb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.591696] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dc8fda-63d8-158a-b11c-9de2d8db3bff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.596712] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd47e3af-47f7-43df-98de-eac781b812ac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.627567] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f490b77-a8e5-4d24-bcfe-8a9e68c1bb4f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.634854] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1aad521-40a2-4787-968a-6d977c027f56 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.647540] env[62814]: DEBUG nova.compute.provider_tree [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1228.931330] env[62814]: DEBUG oslo_vmware.api [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294767, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174909} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.936665] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1228.936862] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1228.937061] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1228.937234] env[62814]: INFO nova.compute.manager [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1228.937465] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1228.937658] env[62814]: DEBUG nova.compute.manager [-] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1228.937748] env[62814]: DEBUG nova.network.neutron [-] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1228.944302] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294768, 'name': Rename_Task, 'duration_secs': 0.146969} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.944568] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1228.944799] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7115f67-99e9-4705-8a89-8ab353ade01b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.952125] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1228.952125] env[62814]: value = "task-4294770" [ 1228.952125] env[62814]: _type = "Task" [ 1228.952125] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1228.959418] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294770, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.004782] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0c97e8c0-c31e-4720-996d-22e04b6a96a8 tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "fed47b28-64fb-4af9-9f8e-97a63afda514" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.524s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1229.041905] env[62814]: DEBUG nova.network.neutron [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance_info_cache with network_info: [{"id": "0235e89b-ec0d-4439-817c-dd76d6987b80", "address": "fa:16:3e:46:9d:6d", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0235e89b-ec", "ovs_interfaceid": "0235e89b-ec0d-4439-817c-dd76d6987b80", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.094979] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dc8fda-63d8-158a-b11c-9de2d8db3bff, 'name': SearchDatastore_Task, 'duration_secs': 0.00866} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.095437] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1229.095521] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] ae9180b8-6caa-44ef-93a8-eb3a7681d224/ae9180b8-6caa-44ef-93a8-eb3a7681d224.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1229.095780] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2836c9a7-7c6b-43af-b52f-f8e1987b581b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.102597] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1229.102597] env[62814]: value = "task-4294771" [ 1229.102597] env[62814]: _type = "Task" [ 1229.102597] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1229.110415] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294771, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.193405] env[62814]: DEBUG nova.scheduler.client.report [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 180 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1229.193405] env[62814]: DEBUG nova.compute.provider_tree [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 180 to 181 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1229.193405] env[62814]: DEBUG nova.compute.provider_tree [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1229.260600] env[62814]: DEBUG nova.compute.manager [req-cc6d1b68-f289-4bfc-ada3-b9c00d732434 req-27cc95fe-c9a1-4d28-b84f-05979947204a service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Received event network-vif-deleted-382b237a-593e-47c2-99c0-18d3918d92a6 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1229.260924] env[62814]: INFO nova.compute.manager [req-cc6d1b68-f289-4bfc-ada3-b9c00d732434 req-27cc95fe-c9a1-4d28-b84f-05979947204a service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Neutron deleted interface 382b237a-593e-47c2-99c0-18d3918d92a6; detaching it from the instance and deleting it from the info cache [ 1229.261207] env[62814]: DEBUG nova.network.neutron [req-cc6d1b68-f289-4bfc-ada3-b9c00d732434 req-27cc95fe-c9a1-4d28-b84f-05979947204a service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.462450] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294770, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.547018] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-b1d88997-e52f-41bd-b1b4-dd096d20d60a" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1229.548298] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77632fb5-1cb4-41bc-8bfe-a060033afa8d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.558161] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a95f2cc-51b3-4e84-bd8b-ba57bf56cde4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.613911] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294771, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1229.715447] env[62814]: DEBUG nova.network.neutron [-] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.764502] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50995c94-e769-4d51-9cc9-194b3e9fa8d8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.775038] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc32ccf0-5e2d-4906-92c8-31c8e2c96ddf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.806022] env[62814]: DEBUG nova.compute.manager [req-cc6d1b68-f289-4bfc-ada3-b9c00d732434 req-27cc95fe-c9a1-4d28-b84f-05979947204a service nova] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Detach interface failed, port_id=382b237a-593e-47c2-99c0-18d3918d92a6, reason: Instance c9b67326-3058-4fa6-a094-b47bf8663444 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1229.962213] env[62814]: DEBUG oslo_vmware.api [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294770, 'name': PowerOnVM_Task, 'duration_secs': 0.727125} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1229.962490] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1229.962740] env[62814]: INFO nova.compute.manager [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Took 8.09 seconds to spawn the instance on the hypervisor. [ 1229.962862] env[62814]: DEBUG nova.compute.manager [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1229.963634] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1917d06-8b52-4493-8252-9013e9e63de0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.114654] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294771, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546376} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.114878] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] ae9180b8-6caa-44ef-93a8-eb3a7681d224/ae9180b8-6caa-44ef-93a8-eb3a7681d224.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1230.115019] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1230.115261] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb4240d4-0ae7-4d2a-abec-d5d66a31f250 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.122133] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1230.122133] env[62814]: value = "task-4294772" [ 1230.122133] env[62814]: _type = "Task" [ 1230.122133] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.131163] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294772, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.207788] env[62814]: DEBUG oslo_concurrency.lockutils [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 3.592s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1230.213848] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.201s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1230.214040] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1230.214269] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1230.216921] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338663fe-806d-4635-baba-d4305b4e1967 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.223670] env[62814]: INFO nova.compute.manager [-] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Took 1.29 seconds to deallocate network for instance. [ 1230.231637] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f963994e-c4dc-4dbb-abf0-92036555e74d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.248089] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a6809a-81f6-4047-90cd-3626bd088137 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.254985] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0439d49f-36af-41be-bc8a-f74ea210cdf4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.286571] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179452MB free_disk=144GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1230.286718] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1230.286939] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1230.404495] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "fed47b28-64fb-4af9-9f8e-97a63afda514" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1230.404709] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "fed47b28-64fb-4af9-9f8e-97a63afda514" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1230.404917] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "fed47b28-64fb-4af9-9f8e-97a63afda514-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1230.405117] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "fed47b28-64fb-4af9-9f8e-97a63afda514-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1230.405320] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "fed47b28-64fb-4af9-9f8e-97a63afda514-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1230.408397] env[62814]: INFO nova.compute.manager [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Terminating instance [ 1230.481480] env[62814]: INFO nova.compute.manager [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Took 16.81 seconds to build instance. [ 1230.632436] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294772, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.159225} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.632743] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1230.633675] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec32bdd-f1bc-4c60-a222-3750eba2d869 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.656867] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] ae9180b8-6caa-44ef-93a8-eb3a7681d224/ae9180b8-6caa-44ef-93a8-eb3a7681d224.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1230.657208] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78b5a6c4-df54-4117-b04b-b2aceb40ee51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.673110] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1230.673110] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af5b3569-4f03-4cc7-8b5f-919a1246a206 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.679320] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1230.679320] env[62814]: value = "task-4294774" [ 1230.679320] env[62814]: _type = "Task" [ 1230.679320] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.680663] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1230.680663] env[62814]: value = "task-4294773" [ 1230.680663] env[62814]: _type = "Task" [ 1230.680663] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.692175] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294774, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.694806] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294773, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.730368] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1230.910910] env[62814]: DEBUG nova.compute.manager [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1230.911172] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1230.912170] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705efbe5-5e35-4db3-bc5a-4371a99e7b3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.920076] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1230.920324] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-686bb82c-ded3-4d60-ba87-4dc576589bf4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.927810] env[62814]: DEBUG oslo_vmware.api [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1230.927810] env[62814]: value = "task-4294775" [ 1230.927810] env[62814]: _type = "Task" [ 1230.927810] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.936962] env[62814]: DEBUG oslo_vmware.api [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294775, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.985718] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c4b8f212-ce23-480f-ae9a-85f1d2854773 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.320s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1231.194237] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294773, 'name': PowerOffVM_Task, 'duration_secs': 0.360549} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.197713] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1231.198443] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1231.198673] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1231.198842] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1231.199057] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1231.199223] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1231.199375] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1231.199596] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1231.199767] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1231.200047] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1231.200120] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1231.200324] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1231.205416] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294774, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.205661] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-366e4de6-7bc2-4927-91da-d09b79ccd0ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.220818] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1231.220818] env[62814]: value = "task-4294776" [ 1231.220818] env[62814]: _type = "Task" [ 1231.220818] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.235052] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294776, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.319134] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 7350d352-9336-40b8-81a6-0a4795d9f8dd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1231.319300] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 1f13051e-e51b-4981-9445-d5420c3c1818 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1231.319453] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1231.387687] env[62814]: DEBUG nova.compute.manager [req-e3f5bfb2-e49f-401f-a475-b55332826bab req-89f200d3-9941-4b6c-9421-d5cedfc41e12 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received event network-changed-9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1231.387687] env[62814]: DEBUG nova.compute.manager [req-e3f5bfb2-e49f-401f-a475-b55332826bab req-89f200d3-9941-4b6c-9421-d5cedfc41e12 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing instance network info cache due to event network-changed-9edabfa8-9760-4ef4-adfb-afc6e2200b50. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1231.387687] env[62814]: DEBUG oslo_concurrency.lockutils [req-e3f5bfb2-e49f-401f-a475-b55332826bab req-89f200d3-9941-4b6c-9421-d5cedfc41e12 service nova] Acquiring lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1231.388015] env[62814]: DEBUG oslo_concurrency.lockutils [req-e3f5bfb2-e49f-401f-a475-b55332826bab req-89f200d3-9941-4b6c-9421-d5cedfc41e12 service nova] Acquired lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1231.388015] env[62814]: DEBUG nova.network.neutron [req-e3f5bfb2-e49f-401f-a475-b55332826bab req-89f200d3-9941-4b6c-9421-d5cedfc41e12 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing network info cache for port 9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1231.438528] env[62814]: DEBUG oslo_vmware.api [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294775, 'name': PowerOffVM_Task, 'duration_secs': 0.23192} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.438812] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1231.439052] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1231.442882] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aa32852f-511b-479a-b3d7-f982ce42e845 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.508174] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1231.508174] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1231.508174] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleting the datastore file [datastore2] fed47b28-64fb-4af9-9f8e-97a63afda514 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1231.508405] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f2b7578-32db-4cb0-99aa-d551ced78adb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.514682] env[62814]: DEBUG oslo_vmware.api [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for the task: (returnval){ [ 1231.514682] env[62814]: value = "task-4294778" [ 1231.514682] env[62814]: _type = "Task" [ 1231.514682] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.523578] env[62814]: DEBUG oslo_vmware.api [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294778, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.691314] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294774, 'name': ReconfigVM_Task, 'duration_secs': 0.528356} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.691771] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Reconfigured VM instance instance-00000075 to attach disk [datastore2] ae9180b8-6caa-44ef-93a8-eb3a7681d224/ae9180b8-6caa-44ef-93a8-eb3a7681d224.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1231.692494] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4cf6fc2c-152e-4112-b912-ce5c39bb440b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.698960] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1231.698960] env[62814]: value = "task-4294780" [ 1231.698960] env[62814]: _type = "Task" [ 1231.698960] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.708264] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294780, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.730266] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294776, 'name': ReconfigVM_Task, 'duration_secs': 0.154753} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.731167] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ced76de-fb99-48a4-baad-68f3b0bb4dce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.757529] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1231.758287] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1231.758630] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1231.758755] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1231.759155] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1231.759155] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1231.759310] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1231.759474] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1231.759652] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1231.759832] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1231.760017] env[62814]: DEBUG nova.virt.hardware [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1231.760952] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e30ea1fc-78bf-49b7-b353-5bfa8a63d5cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.767239] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1231.767239] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52043ca3-1449-45a1-0f7e-4f4d4998e454" [ 1231.767239] env[62814]: _type = "Task" [ 1231.767239] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.771978] env[62814]: INFO nova.compute.manager [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Swapping old allocation on dict_keys(['7136a6f7-3927-4aa1-a4e9-7fcbd9976745']) held by migration 7a60a3b4-92a0-49df-b1ed-bf600dddbef6 for instance [ 1231.783349] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52043ca3-1449-45a1-0f7e-4f4d4998e454, 'name': SearchDatastore_Task, 'duration_secs': 0.007617} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.795497] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1231.796251] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-efacdf4c-1fea-41e4-8824-6f8d0d804a6b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1231.818093] env[62814]: DEBUG nova.scheduler.client.report [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Overwriting current allocation {'allocations': {'7136a6f7-3927-4aa1-a4e9-7fcbd9976745': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 181}}, 'project_id': '513135674ee446d19fa8c667a47138a5', 'user_id': '51348be2493643fb9e5b44f27e5a7a94', 'consumer_generation': 1} on consumer 606e31eb-2349-427f-9c9b-ed9dc5b385f5 {{(pid=62814) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 1231.822920] env[62814]: INFO nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 7a60a3b4-92a0-49df-b1ed-bf600dddbef6 has allocations against this compute host but is not found in the database. [ 1231.823172] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance b1d88997-e52f-41bd-b1b4-dd096d20d60a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1231.823373] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 606e31eb-2349-427f-9c9b-ed9dc5b385f5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1231.823594] env[62814]: WARNING nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance c9b67326-3058-4fa6-a094-b47bf8663444 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1231.823785] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance fed47b28-64fb-4af9-9f8e-97a63afda514 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1231.823968] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance b5a2fb6c-7078-4f3c-b511-014beb96391d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1231.824168] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance ae9180b8-6caa-44ef-93a8-eb3a7681d224 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1231.824458] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1231.824668] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=149GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1231.830040] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1231.830040] env[62814]: value = "task-4294781" [ 1231.830040] env[62814]: _type = "Task" [ 1231.830040] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1231.843013] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294781, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.931503] env[62814]: DEBUG oslo_concurrency.lockutils [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1231.931740] env[62814]: DEBUG oslo_concurrency.lockutils [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1231.931923] env[62814]: DEBUG nova.network.neutron [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1232.005147] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20c3e9a4-751d-45d7-9070-c213b5000a65 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.012837] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a860af2-18ea-4bfa-82b5-505f8c22c084 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.024527] env[62814]: DEBUG oslo_vmware.api [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Task: {'id': task-4294778, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.243214} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.049217] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1232.049439] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1232.049649] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1232.049871] env[62814]: INFO nova.compute.manager [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1232.050143] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1232.052910] env[62814]: DEBUG nova.compute.manager [-] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1232.053017] env[62814]: DEBUG nova.network.neutron [-] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1232.055014] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a800efe1-431f-4070-b299-6e5b1f09debd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.063490] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8285d30-15b2-405b-922d-154d8f14a9a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.078224] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1232.174978] env[62814]: DEBUG nova.network.neutron [req-e3f5bfb2-e49f-401f-a475-b55332826bab req-89f200d3-9941-4b6c-9421-d5cedfc41e12 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updated VIF entry in instance network info cache for port 9edabfa8-9760-4ef4-adfb-afc6e2200b50. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1232.175437] env[62814]: DEBUG nova.network.neutron [req-e3f5bfb2-e49f-401f-a475-b55332826bab req-89f200d3-9941-4b6c-9421-d5cedfc41e12 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [{"id": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "address": "fa:16:3e:ff:ce:8d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9edabfa8-97", "ovs_interfaceid": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.211074] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294780, 'name': Rename_Task, 'duration_secs': 0.237087} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.211467] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1232.211674] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7f031ec-fcb8-4ce0-ab98-3e76ba9bf692 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.219738] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1232.219738] env[62814]: value = "task-4294782" [ 1232.219738] env[62814]: _type = "Task" [ 1232.219738] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.228506] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294782, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.341363] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294781, 'name': ReconfigVM_Task, 'duration_secs': 0.440342} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.341712] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1232.345321] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0761ac4-55d1-4d4b-9091-af48793eef5d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.370735] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] b1d88997-e52f-41bd-b1b4-dd096d20d60a/b1d88997-e52f-41bd-b1b4-dd096d20d60a.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1232.372203] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c3f233f-3e9d-4346-a404-3e5604a32cfd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.394592] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1232.394592] env[62814]: value = "task-4294783" [ 1232.394592] env[62814]: _type = "Task" [ 1232.394592] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.408541] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294783, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.451014] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1232.451383] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1232.603889] env[62814]: ERROR nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [req-b9495abf-d72a-4488-8673-5cfd0eeb0ea8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 7136a6f7-3927-4aa1-a4e9-7fcbd9976745. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b9495abf-d72a-4488-8673-5cfd0eeb0ea8"}]} [ 1232.623543] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1232.638793] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1232.638991] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 144, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1232.651458] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1232.672689] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1232.678593] env[62814]: DEBUG oslo_concurrency.lockutils [req-e3f5bfb2-e49f-401f-a475-b55332826bab req-89f200d3-9941-4b6c-9421-d5cedfc41e12 service nova] Releasing lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1232.689607] env[62814]: DEBUG nova.network.neutron [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance_info_cache with network_info: [{"id": "837d03b5-250f-44d2-bafb-31184267117d", "address": "fa:16:3e:24:ee:bf", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap837d03b5-25", "ovs_interfaceid": "837d03b5-250f-44d2-bafb-31184267117d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.730559] env[62814]: DEBUG oslo_vmware.api [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294782, 'name': PowerOnVM_Task, 'duration_secs': 0.492195} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.733323] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1232.733540] env[62814]: INFO nova.compute.manager [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Took 8.59 seconds to spawn the instance on the hypervisor. [ 1232.733723] env[62814]: DEBUG nova.compute.manager [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1232.734776] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af031827-ce20-423d-b5e1-1a8ce7c71d0a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.813046] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1eb787a-3f2b-47cd-8608-4c81dbfa5274 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.820857] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2275577a-4135-4818-b711-c36c658efb7f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.852151] env[62814]: DEBUG nova.network.neutron [-] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.854147] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80218831-cabb-49b5-96ce-a1b8185d58dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.862695] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85785836-bcdb-493f-890b-38866ae18d63 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.876997] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1232.904692] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294783, 'name': ReconfigVM_Task, 'duration_secs': 0.353654} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.904970] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfigured VM instance instance-00000067 to attach disk [datastore2] b1d88997-e52f-41bd-b1b4-dd096d20d60a/b1d88997-e52f-41bd-b1b4-dd096d20d60a.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1232.905922] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076c6556-31a7-449f-8491-3ef8ccbf8815 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.927941] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4343ea-abb4-4baf-b3ec-1260f5ec95e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.948677] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36006b9e-75ef-410d-b6ce-9344e22bed56 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.972021] env[62814]: DEBUG nova.compute.manager [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1232.974316] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550e9240-6b7a-420b-b97d-b3758ffe8941 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.981998] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1232.982206] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7cf64d7d-e46e-48eb-a14e-d2d1408e31e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.988316] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1232.988316] env[62814]: value = "task-4294784" [ 1232.988316] env[62814]: _type = "Task" [ 1232.988316] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.995785] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294784, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.194628] env[62814]: DEBUG oslo_concurrency.lockutils [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-606e31eb-2349-427f-9c9b-ed9dc5b385f5" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1233.195026] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1233.195339] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2cbc85e9-07e6-4da5-a711-56dac743448e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.203818] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1233.203818] env[62814]: value = "task-4294785" [ 1233.203818] env[62814]: _type = "Task" [ 1233.203818] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.211182] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294785, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.254952] env[62814]: INFO nova.compute.manager [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Took 15.89 seconds to build instance. [ 1233.357624] env[62814]: INFO nova.compute.manager [-] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Took 1.30 seconds to deallocate network for instance. [ 1233.413518] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updated inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with generation 183 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 1233.413791] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 generation from 183 to 184 during operation: update_inventory {{(pid=62814) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1233.414090] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1233.419414] env[62814]: DEBUG nova.compute.manager [req-87288e75-1ac2-4b67-abef-cf7b6ac82912 req-5474339d-b132-408c-9e28-fdf02c237be8 service nova] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Received event network-vif-deleted-0c147c05-6c14-4d99-ac3f-33be08394039 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1233.493142] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1233.499362] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294784, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.714596] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294785, 'name': PowerOffVM_Task, 'duration_secs': 0.194993} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.714796] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1233.715499] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1233.715720] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1233.715878] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1233.716075] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1233.716578] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1233.716578] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1233.716578] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1233.716737] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1233.716876] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1233.717051] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1233.717237] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1233.722200] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1317f222-4542-4127-b331-dc6a1d607e32 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.738162] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1233.738162] env[62814]: value = "task-4294786" [ 1233.738162] env[62814]: _type = "Task" [ 1233.738162] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1233.747963] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294786, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.756563] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d9bce64b-be04-4e33-9e57-936abcff39e7 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.401s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1233.864095] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1233.921170] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1233.921501] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.634s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1233.921863] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.192s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1233.922079] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1233.924445] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.431s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1233.926269] env[62814]: INFO nova.compute.claims [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1233.944037] env[62814]: INFO nova.scheduler.client.report [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted allocations for instance c9b67326-3058-4fa6-a094-b47bf8663444 [ 1233.998166] env[62814]: DEBUG oslo_vmware.api [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294784, 'name': PowerOnVM_Task, 'duration_secs': 0.629155} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.998452] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1234.247406] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294786, 'name': ReconfigVM_Task, 'duration_secs': 0.136794} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.248283] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bf2d01-ac67-4ffd-960d-d31d605bf574 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.265969] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1234.266191] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1234.266785] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1234.266785] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1234.266785] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1234.267065] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1234.267065] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1234.267182] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1234.267355] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1234.267524] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1234.267701] env[62814]: DEBUG nova.virt.hardware [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1234.269633] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb500e3f-7b03-40a8-ac43-e532ec7d6cf8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.279436] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1234.279436] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272f7c9-e369-432a-81f4-566397f8c0e2" [ 1234.279436] env[62814]: _type = "Task" [ 1234.279436] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.287759] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272f7c9-e369-432a-81f4-566397f8c0e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.451579] env[62814]: DEBUG oslo_concurrency.lockutils [None req-f2dfeb1c-0fd7-4251-8a61-60865508d9ac tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "c9b67326-3058-4fa6-a094-b47bf8663444" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.626s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1234.784908] env[62814]: DEBUG nova.compute.manager [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Stashing vm_state: active {{(pid=62814) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1234.793926] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5272f7c9-e369-432a-81f4-566397f8c0e2, 'name': SearchDatastore_Task, 'duration_secs': 0.017927} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.799992] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1234.800279] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ca04502-766c-4e8e-8eb9-d7926b1e1c45 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.818099] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1234.818099] env[62814]: value = "task-4294787" [ 1234.818099] env[62814]: _type = "Task" [ 1234.818099] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.825933] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294787, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.925090] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1234.925392] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1234.925581] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1235.055482] env[62814]: INFO nova.compute.manager [None req-3780241a-aebc-49d7-ba1e-57392e037c63 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance to original state: 'active' [ 1235.096155] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5c2527-7f9b-4e53-9dac-a45711749d27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.105092] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db664c6-03f4-4054-9838-534efef57739 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.138021] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6391cd0f-6fe5-437d-88f9-561afb3c5248 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.146276] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a0a24d5-3dc6-4ee2-9bfc-06c850568479 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.161255] env[62814]: DEBUG nova.compute.provider_tree [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1235.307823] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1235.331095] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294787, 'name': ReconfigVM_Task, 'duration_secs': 0.19709} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.331095] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1235.331095] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6170ea59-8609-4a42-8a81-b301e6ad527c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.354648] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 606e31eb-2349-427f-9c9b-ed9dc5b385f5/606e31eb-2349-427f-9c9b-ed9dc5b385f5.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1235.357050] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4946a269-c978-47d1-a986-bf16a9daf9a2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.379182] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1235.379182] env[62814]: value = "task-4294788" [ 1235.379182] env[62814]: _type = "Task" [ 1235.379182] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1235.389955] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294788, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.665234] env[62814]: DEBUG nova.scheduler.client.report [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1235.873754] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "7350d352-9336-40b8-81a6-0a4795d9f8dd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1235.874018] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "7350d352-9336-40b8-81a6-0a4795d9f8dd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1235.875652] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "7350d352-9336-40b8-81a6-0a4795d9f8dd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1235.875652] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "7350d352-9336-40b8-81a6-0a4795d9f8dd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1235.875652] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "7350d352-9336-40b8-81a6-0a4795d9f8dd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1235.877553] env[62814]: INFO nova.compute.manager [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Terminating instance [ 1235.890665] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294788, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.171328] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.247s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1236.171871] env[62814]: DEBUG nova.compute.manager [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1236.175955] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.312s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1236.176283] env[62814]: DEBUG nova.objects.instance [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lazy-loading 'resources' on Instance uuid fed47b28-64fb-4af9-9f8e-97a63afda514 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1236.373831] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1236.374805] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1236.374904] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1236.375176] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1236.375238] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1236.377451] env[62814]: INFO nova.compute.manager [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Terminating instance [ 1236.380432] env[62814]: DEBUG nova.compute.manager [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1236.380591] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1236.381607] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46dc963-3d57-4cde-ae6f-345dfe07d115 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.393807] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294788, 'name': ReconfigVM_Task, 'duration_secs': 0.818116} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.396311] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 606e31eb-2349-427f-9c9b-ed9dc5b385f5/606e31eb-2349-427f-9c9b-ed9dc5b385f5.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1236.397054] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1236.397741] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f307d28-3343-4b27-a798-83ae58e54c72 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.400078] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0985402a-3cdf-43c4-82b1-ad9ffc58e621 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.420192] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16cb7b95-7d7f-4cfb-917f-46dd1f475d6e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.422863] env[62814]: DEBUG oslo_vmware.api [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1236.422863] env[62814]: value = "task-4294789" [ 1236.422863] env[62814]: _type = "Task" [ 1236.422863] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.444129] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba3eac7-5fe4-4abb-bf17-652729ef1cea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.450374] env[62814]: DEBUG oslo_vmware.api [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294789, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.467320] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92befba-1258-4947-8911-4b316b3e7a4f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.474462] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1236.474729] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e87d9241-3252-41c8-b6b2-aebc91f3dd7d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.481302] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1236.481302] env[62814]: value = "task-4294790" [ 1236.481302] env[62814]: _type = "Task" [ 1236.481302] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.489181] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294790, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.679086] env[62814]: DEBUG nova.compute.utils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1236.681078] env[62814]: DEBUG nova.compute.manager [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1236.681078] env[62814]: DEBUG nova.network.neutron [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1236.734043] env[62814]: DEBUG nova.policy [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5779eaa60ec44b2b80660e521b598eb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10018060297c409d82935ad626cc54c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1236.844773] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3c8f1bb-c7bc-4e1c-9d92-a6f93637fa7f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.853146] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5503adf9-60b9-4e65-9a06-03a526e43e89 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.885513] env[62814]: DEBUG nova.compute.manager [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1236.885870] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1236.886339] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7731fa15-fdc5-468f-b5d9-966ea8c1865f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.888575] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c92baff-2714-4cc2-813d-d8343d7d46e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.897474] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f164cb2-215d-4b71-a1fc-65a5d2dde694 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.903660] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1236.903660] env[62814]: value = "task-4294791" [ 1236.903660] env[62814]: _type = "Task" [ 1236.903660] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1236.920370] env[62814]: DEBUG nova.compute.provider_tree [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1236.926234] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294791, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1236.934386] env[62814]: DEBUG oslo_vmware.api [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294789, 'name': PowerOffVM_Task, 'duration_secs': 0.304462} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.935481] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1236.935693] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1236.935952] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-91d10173-d9f1-46f5-b636-8fe49fc8ab6f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1236.992367] env[62814]: DEBUG oslo_vmware.api [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294790, 'name': PowerOnVM_Task, 'duration_secs': 0.380758} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1236.993761] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1237.004187] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1237.004187] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1237.004374] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleting the datastore file [datastore2] 7350d352-9336-40b8-81a6-0a4795d9f8dd {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1237.004592] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4710a065-c28c-4c53-8cac-7d988d94299f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.012156] env[62814]: DEBUG oslo_vmware.api [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for the task: (returnval){ [ 1237.012156] env[62814]: value = "task-4294793" [ 1237.012156] env[62814]: _type = "Task" [ 1237.012156] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.021947] env[62814]: DEBUG oslo_vmware.api [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294793, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.101964] env[62814]: DEBUG nova.network.neutron [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Successfully created port: 90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1237.186787] env[62814]: DEBUG nova.compute.manager [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1237.414103] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294791, 'name': PowerOffVM_Task, 'duration_secs': 0.292885} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.414443] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1237.414651] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1237.414851] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845857', 'volume_id': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'name': 'volume-808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'b1d88997-e52f-41bd-b1b4-dd096d20d60a', 'attached_at': '2025-06-21T05:44:38.000000', 'detached_at': '', 'volume_id': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'serial': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1237.415610] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a95ef9-dcbe-4a0c-a578-c264ea3a412c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.437149] env[62814]: DEBUG nova.scheduler.client.report [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1237.441352] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30e395f1-7c23-4afe-83c6-420fd504df15 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.449242] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dad74fd-727c-43ed-a709-3c7b30412435 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.469914] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a77ff9-2262-471e-a11c-2b2de518383e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.486564] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] The volume has not been displaced from its original location: [datastore1] volume-808c711a-a0ee-42d1-96d6-cb028b2a9fe7/volume-808c711a-a0ee-42d1-96d6-cb028b2a9fe7.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1237.491703] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfiguring VM instance instance-00000067 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1237.492010] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0fb90aa-09bf-4b23-800e-81f97eb2d44c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1237.513356] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1237.513356] env[62814]: value = "task-4294794" [ 1237.513356] env[62814]: _type = "Task" [ 1237.513356] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1237.523986] env[62814]: DEBUG oslo_vmware.api [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Task: {'id': task-4294793, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.324469} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1237.527090] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1237.527297] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1237.527497] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1237.527675] env[62814]: INFO nova.compute.manager [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1237.527909] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1237.528118] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294794, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1237.528326] env[62814]: DEBUG nova.compute.manager [-] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1237.528423] env[62814]: DEBUG nova.network.neutron [-] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1237.875783] env[62814]: DEBUG nova.compute.manager [req-ec493148-ee01-494e-b3a0-9101ac828f8f req-d8cc9e4d-41eb-466e-a550-0130046e6c37 service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Received event network-vif-deleted-d5a713c4-180b-4ecd-9c1d-e2ad28f526ad {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1237.876040] env[62814]: INFO nova.compute.manager [req-ec493148-ee01-494e-b3a0-9101ac828f8f req-d8cc9e4d-41eb-466e-a550-0130046e6c37 service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Neutron deleted interface d5a713c4-180b-4ecd-9c1d-e2ad28f526ad; detaching it from the instance and deleting it from the info cache [ 1237.876169] env[62814]: DEBUG nova.network.neutron [req-ec493148-ee01-494e-b3a0-9101ac828f8f req-d8cc9e4d-41eb-466e-a550-0130046e6c37 service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1237.945996] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1237.948234] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.641s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1237.968334] env[62814]: INFO nova.scheduler.client.report [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Deleted allocations for instance fed47b28-64fb-4af9-9f8e-97a63afda514 [ 1238.011651] env[62814]: INFO nova.compute.manager [None req-04e9fe43-dda4-45ad-ab18-c0850307ea07 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance to original state: 'active' [ 1238.026167] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294794, 'name': ReconfigVM_Task, 'duration_secs': 0.27156} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.026455] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Reconfigured VM instance instance-00000067 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1238.031548] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94db5f8a-c822-464f-9047-be91fb225d4e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.048448] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1238.048448] env[62814]: value = "task-4294795" [ 1238.048448] env[62814]: _type = "Task" [ 1238.048448] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.055995] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294795, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.197927] env[62814]: DEBUG nova.compute.manager [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1238.223600] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1238.223843] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1238.224007] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1238.224202] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1238.224492] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1238.224734] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1238.224965] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1238.225155] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1238.225350] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1238.225558] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1238.225749] env[62814]: DEBUG nova.virt.hardware [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1238.226808] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-098da124-f06e-4356-a031-6c6aefafa1d3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.235133] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3eab896-7549-46f8-a73b-7c422f160a05 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.325966] env[62814]: DEBUG nova.network.neutron [-] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1238.378564] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac25afd2-c629-4692-a21e-4538743374aa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.387711] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4af11c04-275c-4f48-814b-9b6edd72ed12 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.415771] env[62814]: DEBUG nova.compute.manager [req-ec493148-ee01-494e-b3a0-9101ac828f8f req-d8cc9e4d-41eb-466e-a550-0130046e6c37 service nova] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Detach interface failed, port_id=d5a713c4-180b-4ecd-9c1d-e2ad28f526ad, reason: Instance 7350d352-9336-40b8-81a6-0a4795d9f8dd could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1238.452843] env[62814]: INFO nova.compute.claims [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1238.475408] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e9b184fa-8483-4f81-91df-05e495fc8f5d tempest-ServerDiskConfigTestJSON-1215934053 tempest-ServerDiskConfigTestJSON-1215934053-project-member] Lock "fed47b28-64fb-4af9-9f8e-97a63afda514" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.071s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1238.545872] env[62814]: DEBUG nova.compute.manager [req-cbe68842-c2b8-4b24-be9c-5bcd0c83f2e1 req-ab9deeed-cc02-49fa-aefc-03292a21bea1 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Received event network-vif-plugged-90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1238.545872] env[62814]: DEBUG oslo_concurrency.lockutils [req-cbe68842-c2b8-4b24-be9c-5bcd0c83f2e1 req-ab9deeed-cc02-49fa-aefc-03292a21bea1 service nova] Acquiring lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1238.545985] env[62814]: DEBUG oslo_concurrency.lockutils [req-cbe68842-c2b8-4b24-be9c-5bcd0c83f2e1 req-ab9deeed-cc02-49fa-aefc-03292a21bea1 service nova] Lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1238.546098] env[62814]: DEBUG oslo_concurrency.lockutils [req-cbe68842-c2b8-4b24-be9c-5bcd0c83f2e1 req-ab9deeed-cc02-49fa-aefc-03292a21bea1 service nova] Lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1238.546263] env[62814]: DEBUG nova.compute.manager [req-cbe68842-c2b8-4b24-be9c-5bcd0c83f2e1 req-ab9deeed-cc02-49fa-aefc-03292a21bea1 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] No waiting events found dispatching network-vif-plugged-90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1238.546620] env[62814]: WARNING nova.compute.manager [req-cbe68842-c2b8-4b24-be9c-5bcd0c83f2e1 req-ab9deeed-cc02-49fa-aefc-03292a21bea1 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Received unexpected event network-vif-plugged-90c053d0-8e56-42b1-89da-b7acd45ac0fa for instance with vm_state building and task_state spawning. [ 1238.557782] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294795, 'name': ReconfigVM_Task, 'duration_secs': 0.391458} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1238.558378] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845857', 'volume_id': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'name': 'volume-808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'b1d88997-e52f-41bd-b1b4-dd096d20d60a', 'attached_at': '2025-06-21T05:44:38.000000', 'detached_at': '', 'volume_id': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7', 'serial': '808c711a-a0ee-42d1-96d6-cb028b2a9fe7'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1238.558606] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1238.559402] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5461ec80-5e32-4974-a877-eaedaf369867 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.566676] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1238.566930] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-472d4ba9-43b0-4ab1-b606-88e3fa30e25e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.625209] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1238.625443] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1238.625626] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleting the datastore file [datastore2] b1d88997-e52f-41bd-b1b4-dd096d20d60a {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1238.625880] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4aa1273a-7ece-406f-8180-0b57fc788973 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.633025] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1238.633025] env[62814]: value = "task-4294797" [ 1238.633025] env[62814]: _type = "Task" [ 1238.633025] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.641048] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294797, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1238.675800] env[62814]: DEBUG nova.network.neutron [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Successfully updated port: 90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1238.829023] env[62814]: INFO nova.compute.manager [-] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Took 1.30 seconds to deallocate network for instance. [ 1238.959239] env[62814]: INFO nova.compute.resource_tracker [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating resource usage from migration e0d9dda5-4f55-449b-b60e-95c12717178c [ 1239.094015] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129ae4da-824a-4abb-ae52-f01b5465fba5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.104221] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19cab8ea-7f46-44c3-8cba-7c3475b352e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.137795] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f073c3c3-c585-4a8b-aa9c-a7bae670ad9d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.145097] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294797, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.149339] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96811cf7-a1a2-42f8-bffb-f53ae5bbf898 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.164923] env[62814]: DEBUG nova.compute.provider_tree [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1239.177540] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1239.177741] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1239.177858] env[62814]: DEBUG nova.network.neutron [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1239.336654] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1239.644428] env[62814]: DEBUG oslo_vmware.api [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294797, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.527333} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.644730] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1239.644902] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1239.645128] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1239.645361] env[62814]: INFO nova.compute.manager [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Took 2.76 seconds to destroy the instance on the hypervisor. [ 1239.645613] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1239.645811] env[62814]: DEBUG nova.compute.manager [-] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1239.645913] env[62814]: DEBUG nova.network.neutron [-] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1239.668232] env[62814]: DEBUG nova.scheduler.client.report [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1239.740342] env[62814]: DEBUG nova.network.neutron [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1239.933760] env[62814]: DEBUG nova.network.neutron [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updating instance_info_cache with network_info: [{"id": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "address": "fa:16:3e:e7:24:4d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c053d0-8e", "ovs_interfaceid": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1240.038268] env[62814]: DEBUG oslo_concurrency.lockutils [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1240.038530] env[62814]: DEBUG oslo_concurrency.lockutils [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1240.038738] env[62814]: DEBUG oslo_concurrency.lockutils [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1240.038917] env[62814]: DEBUG oslo_concurrency.lockutils [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1240.039095] env[62814]: DEBUG oslo_concurrency.lockutils [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1240.043332] env[62814]: INFO nova.compute.manager [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Terminating instance [ 1240.173294] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.225s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1240.173526] env[62814]: INFO nova.compute.manager [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Migrating [ 1240.185123] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.849s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1240.185448] env[62814]: DEBUG nova.objects.instance [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lazy-loading 'resources' on Instance uuid 7350d352-9336-40b8-81a6-0a4795d9f8dd {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1240.439629] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1240.440550] env[62814]: DEBUG nova.compute.manager [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Instance network_info: |[{"id": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "address": "fa:16:3e:e7:24:4d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c053d0-8e", "ovs_interfaceid": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1240.440795] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e7:24:4d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe99da4f-5630-4afd-918b-b327193d8489', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90c053d0-8e56-42b1-89da-b7acd45ac0fa', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1240.449449] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1240.449719] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1240.449981] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8768156d-abe4-4764-9ca6-99a285ab8f73 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.472613] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1240.472613] env[62814]: value = "task-4294798" [ 1240.472613] env[62814]: _type = "Task" [ 1240.472613] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.483161] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294798, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.549020] env[62814]: DEBUG nova.compute.manager [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1240.549020] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1240.549020] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6900a197-c3c9-437c-a8c2-ca40a21bbc75 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.557022] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1240.557022] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee7305a6-2ba2-4e49-ad82-4c3ab7581d5b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.563494] env[62814]: DEBUG oslo_vmware.api [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1240.563494] env[62814]: value = "task-4294799" [ 1240.563494] env[62814]: _type = "Task" [ 1240.563494] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.573774] env[62814]: DEBUG oslo_vmware.api [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294799, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.591407] env[62814]: DEBUG nova.compute.manager [req-d69a375a-9af3-44f6-8ab6-06c404a05342 req-025e2256-0b59-4a7f-87b7-793b3b1c573d service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Received event network-changed-90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1240.591407] env[62814]: DEBUG nova.compute.manager [req-d69a375a-9af3-44f6-8ab6-06c404a05342 req-025e2256-0b59-4a7f-87b7-793b3b1c573d service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Refreshing instance network info cache due to event network-changed-90c053d0-8e56-42b1-89da-b7acd45ac0fa. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1240.591407] env[62814]: DEBUG oslo_concurrency.lockutils [req-d69a375a-9af3-44f6-8ab6-06c404a05342 req-025e2256-0b59-4a7f-87b7-793b3b1c573d service nova] Acquiring lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.591772] env[62814]: DEBUG oslo_concurrency.lockutils [req-d69a375a-9af3-44f6-8ab6-06c404a05342 req-025e2256-0b59-4a7f-87b7-793b3b1c573d service nova] Acquired lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1240.592160] env[62814]: DEBUG nova.network.neutron [req-d69a375a-9af3-44f6-8ab6-06c404a05342 req-025e2256-0b59-4a7f-87b7-793b3b1c573d service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Refreshing network info cache for port 90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1240.694102] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.694102] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1240.694102] env[62814]: DEBUG nova.network.neutron [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1240.850140] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91008c6b-ffe2-48dd-a6ea-871148c9f695 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.858162] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c197ef-7df1-41f4-9542-5776394b730d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.896064] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f002e5d2-3508-4eeb-b07d-ec4feb434a72 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.903939] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c2b4a9-99a5-4714-b09f-1d7f4ed72463 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.918360] env[62814]: DEBUG nova.compute.provider_tree [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1240.986525] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294798, 'name': CreateVM_Task, 'duration_secs': 0.434209} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.986525] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1240.986525] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.986525] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1240.986525] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1240.986525] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d72e72b-b7dc-42b5-8f3a-06d895b434b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.990318] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1240.990318] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5257eff0-4b2a-85e0-4fa9-1952172e56e0" [ 1240.990318] env[62814]: _type = "Task" [ 1240.990318] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.000530] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5257eff0-4b2a-85e0-4fa9-1952172e56e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.072966] env[62814]: DEBUG oslo_vmware.api [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294799, 'name': PowerOffVM_Task, 'duration_secs': 0.23355} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.073262] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1241.073509] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1241.073775] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca486a0f-901c-4018-82f4-0264c3ab1759 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.139471] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1241.139643] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1241.139830] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleting the datastore file [datastore2] 606e31eb-2349-427f-9c9b-ed9dc5b385f5 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1241.140119] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eec96f77-0402-466d-b563-5dccd2af3d1a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.147317] env[62814]: DEBUG oslo_vmware.api [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1241.147317] env[62814]: value = "task-4294801" [ 1241.147317] env[62814]: _type = "Task" [ 1241.147317] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.157282] env[62814]: DEBUG oslo_vmware.api [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294801, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.209644] env[62814]: DEBUG nova.network.neutron [-] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1241.338930] env[62814]: DEBUG nova.network.neutron [req-d69a375a-9af3-44f6-8ab6-06c404a05342 req-025e2256-0b59-4a7f-87b7-793b3b1c573d service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updated VIF entry in instance network info cache for port 90c053d0-8e56-42b1-89da-b7acd45ac0fa. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1241.338930] env[62814]: DEBUG nova.network.neutron [req-d69a375a-9af3-44f6-8ab6-06c404a05342 req-025e2256-0b59-4a7f-87b7-793b3b1c573d service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updating instance_info_cache with network_info: [{"id": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "address": "fa:16:3e:e7:24:4d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c053d0-8e", "ovs_interfaceid": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1241.421278] env[62814]: DEBUG nova.scheduler.client.report [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1241.504647] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5257eff0-4b2a-85e0-4fa9-1952172e56e0, 'name': SearchDatastore_Task, 'duration_secs': 0.035944} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.505123] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1241.505266] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1241.506527] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.506527] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1241.507234] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1241.507955] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-10a87e21-4f8c-47e5-8f29-7e26a10fd889 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.518125] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1241.519168] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1241.520117] env[62814]: DEBUG nova.network.neutron [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance_info_cache with network_info: [{"id": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "address": "fa:16:3e:94:da:3d", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d17bc61-a2", "ovs_interfaceid": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1241.521335] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8f40629-31ee-4a27-a343-e2018708edc8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.527502] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1241.527502] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c84980-7a45-c420-a48f-2555b0684236" [ 1241.527502] env[62814]: _type = "Task" [ 1241.527502] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.536546] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c84980-7a45-c420-a48f-2555b0684236, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.657443] env[62814]: DEBUG oslo_vmware.api [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294801, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196833} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.658105] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1241.658105] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1241.658105] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1241.658284] env[62814]: INFO nova.compute.manager [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1241.659122] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1241.659122] env[62814]: DEBUG nova.compute.manager [-] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1241.659122] env[62814]: DEBUG nova.network.neutron [-] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1241.712644] env[62814]: INFO nova.compute.manager [-] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Took 2.07 seconds to deallocate network for instance. [ 1241.843813] env[62814]: DEBUG oslo_concurrency.lockutils [req-d69a375a-9af3-44f6-8ab6-06c404a05342 req-025e2256-0b59-4a7f-87b7-793b3b1c573d service nova] Releasing lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1241.926621] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1241.952410] env[62814]: INFO nova.scheduler.client.report [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Deleted allocations for instance 7350d352-9336-40b8-81a6-0a4795d9f8dd [ 1242.025232] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1242.051295] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c84980-7a45-c420-a48f-2555b0684236, 'name': SearchDatastore_Task, 'duration_secs': 0.010565} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.052904] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7659fdb-35e9-4191-8214-812b0900c912 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.063573] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1242.063573] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5263c105-3102-97a3-938d-30d0fc408297" [ 1242.063573] env[62814]: _type = "Task" [ 1242.063573] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.075310] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5263c105-3102-97a3-938d-30d0fc408297, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.267026] env[62814]: INFO nova.compute.manager [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Took 0.55 seconds to detach 1 volumes for instance. [ 1242.302606] env[62814]: DEBUG oslo_concurrency.lockutils [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "1f13051e-e51b-4981-9445-d5420c3c1818" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1242.303160] env[62814]: DEBUG oslo_concurrency.lockutils [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.462066] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d29778e0-efb2-46e4-9583-6d7985686cda tempest-ServersTestJSON-1829400378 tempest-ServersTestJSON-1829400378-project-member] Lock "7350d352-9336-40b8-81a6-0a4795d9f8dd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.587s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1242.578937] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5263c105-3102-97a3-938d-30d0fc408297, 'name': SearchDatastore_Task, 'duration_secs': 0.010838} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.579391] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1242.582025] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 7b82801d-7a0f-4baf-ad2e-566174d3eb35/7b82801d-7a0f-4baf-ad2e-566174d3eb35.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1242.582422] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7be4b28b-c804-4264-a8a7-71f2c1e7bdcd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.591338] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1242.591338] env[62814]: value = "task-4294802" [ 1242.591338] env[62814]: _type = "Task" [ 1242.591338] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.601986] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294802, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.633258] env[62814]: DEBUG nova.compute.manager [req-edabbfa2-e517-4542-bd8a-99e81ab86690 req-7f4709b6-1852-4340-bbc8-3c9055ec1c53 service nova] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Received event network-vif-deleted-0235e89b-ec0d-4439-817c-dd76d6987b80 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1242.774478] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1242.774899] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1242.775307] env[62814]: DEBUG nova.objects.instance [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'resources' on Instance uuid b1d88997-e52f-41bd-b1b4-dd096d20d60a {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1242.798902] env[62814]: DEBUG nova.compute.manager [req-29b53352-022a-496e-baa3-0262e4ead084 req-0bbafe60-d7ff-40ec-8998-f08e7ca078d5 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Received event network-vif-deleted-837d03b5-250f-44d2-bafb-31184267117d {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1242.799434] env[62814]: INFO nova.compute.manager [req-29b53352-022a-496e-baa3-0262e4ead084 req-0bbafe60-d7ff-40ec-8998-f08e7ca078d5 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Neutron deleted interface 837d03b5-250f-44d2-bafb-31184267117d; detaching it from the instance and deleting it from the info cache [ 1242.799434] env[62814]: DEBUG nova.network.neutron [req-29b53352-022a-496e-baa3-0262e4ead084 req-0bbafe60-d7ff-40ec-8998-f08e7ca078d5 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.807217] env[62814]: INFO nova.compute.manager [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Detaching volume 3e53f486-225a-44b7-bf8a-b91c6ac86fda [ 1242.856032] env[62814]: INFO nova.virt.block_device [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Attempting to driver detach volume 3e53f486-225a-44b7-bf8a-b91c6ac86fda from mountpoint /dev/sdb [ 1242.856032] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1242.856032] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845855', 'volume_id': '3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'name': 'volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '1f13051e-e51b-4981-9445-d5420c3c1818', 'attached_at': '', 'detached_at': '', 'volume_id': '3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'serial': '3e53f486-225a-44b7-bf8a-b91c6ac86fda'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1242.856686] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2032b2ff-b3c8-4b3b-b361-b3a0e2c1a985 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.886461] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f13d6c9-d1c0-4e4f-ab08-4a9993a96a56 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.893457] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4342da9-bd24-43e4-b12d-06ed6ce6da2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.914499] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ecc05fc-c8b8-4d11-8e37-7288c165541e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.930024] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] The volume has not been displaced from its original location: [datastore1] volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda/volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1242.935766] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1242.936031] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-539d0fc1-529f-48e5-863b-b080f7e12898 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.954904] env[62814]: DEBUG oslo_vmware.api [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1242.954904] env[62814]: value = "task-4294803" [ 1242.954904] env[62814]: _type = "Task" [ 1242.954904] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.965772] env[62814]: DEBUG oslo_vmware.api [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294803, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.101858] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294802, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.262545] env[62814]: DEBUG nova.network.neutron [-] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1243.302317] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91470f80-8fa3-48ca-9307-aae8640fd974 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.312125] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3bafba2-33ac-48ef-be57-7cc74a5dfd88 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.348262] env[62814]: DEBUG nova.compute.manager [req-29b53352-022a-496e-baa3-0262e4ead084 req-0bbafe60-d7ff-40ec-8998-f08e7ca078d5 service nova] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Detach interface failed, port_id=837d03b5-250f-44d2-bafb-31184267117d, reason: Instance 606e31eb-2349-427f-9c9b-ed9dc5b385f5 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1243.429319] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b1d2f8-d073-41b1-8fa7-34cf79917da1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.436560] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df3bc4b-b8ed-450d-953c-b6120355cc1a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.470418] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f98556f-87e4-488c-a6a5-ffc9103dc85b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.480965] env[62814]: DEBUG oslo_vmware.api [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294803, 'name': ReconfigVM_Task, 'duration_secs': 0.313331} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.481280] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1243.485897] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b5f7d79-212d-4ad0-a4ee-ee02c16be648 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.496581] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc3cbf8-4c34-4c3c-97d1-d4569c6381a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.510615] env[62814]: DEBUG nova.compute.provider_tree [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1243.516019] env[62814]: DEBUG oslo_vmware.api [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1243.516019] env[62814]: value = "task-4294804" [ 1243.516019] env[62814]: _type = "Task" [ 1243.516019] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.522016] env[62814]: DEBUG oslo_vmware.api [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294804, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.550769] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f883bf-4a92-43be-9b80-8e3b5236567b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.569470] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance 'ae9180b8-6caa-44ef-93a8-eb3a7681d224' progress to 0 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1243.601520] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294802, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511325} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.602162] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 7b82801d-7a0f-4baf-ad2e-566174d3eb35/7b82801d-7a0f-4baf-ad2e-566174d3eb35.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1243.602162] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1243.602345] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-65320975-3198-459f-8075-daa7e0dd2251 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.609503] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1243.609503] env[62814]: value = "task-4294805" [ 1243.609503] env[62814]: _type = "Task" [ 1243.609503] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.617106] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294805, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.765516] env[62814]: INFO nova.compute.manager [-] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Took 2.11 seconds to deallocate network for instance. [ 1244.016365] env[62814]: DEBUG nova.scheduler.client.report [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1244.028732] env[62814]: DEBUG oslo_vmware.api [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294804, 'name': ReconfigVM_Task, 'duration_secs': 0.141873} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.029047] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845855', 'volume_id': '3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'name': 'volume-3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attached', 'instance': '1f13051e-e51b-4981-9445-d5420c3c1818', 'attached_at': '', 'detached_at': '', 'volume_id': '3e53f486-225a-44b7-bf8a-b91c6ac86fda', 'serial': '3e53f486-225a-44b7-bf8a-b91c6ac86fda'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1244.077642] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1244.077951] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28fae60f-c4a4-4366-b591-53a5c267045d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.085126] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1244.085126] env[62814]: value = "task-4294806" [ 1244.085126] env[62814]: _type = "Task" [ 1244.085126] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.094309] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294806, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.119813] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294805, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068249} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.121067] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1244.121196] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32cbc197-9bf8-456f-94c8-d2db23769a0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.144748] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Reconfiguring VM instance instance-00000076 to attach disk [datastore2] 7b82801d-7a0f-4baf-ad2e-566174d3eb35/7b82801d-7a0f-4baf-ad2e-566174d3eb35.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1244.145075] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89b07072-800d-44db-9dce-0479c40639db {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.169888] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1244.169888] env[62814]: value = "task-4294807" [ 1244.169888] env[62814]: _type = "Task" [ 1244.169888] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.177888] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294807, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.271634] env[62814]: DEBUG oslo_concurrency.lockutils [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1244.525395] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.750s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1244.527821] env[62814]: DEBUG oslo_concurrency.lockutils [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.257s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1244.528056] env[62814]: DEBUG nova.objects.instance [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'resources' on Instance uuid 606e31eb-2349-427f-9c9b-ed9dc5b385f5 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1244.550691] env[62814]: INFO nova.scheduler.client.report [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleted allocations for instance b1d88997-e52f-41bd-b1b4-dd096d20d60a [ 1244.578472] env[62814]: DEBUG nova.objects.instance [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'flavor' on Instance uuid 1f13051e-e51b-4981-9445-d5420c3c1818 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1244.595296] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294806, 'name': PowerOffVM_Task, 'duration_secs': 0.185142} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.595622] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1244.596082] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance 'ae9180b8-6caa-44ef-93a8-eb3a7681d224' progress to 17 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1244.679623] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294807, 'name': ReconfigVM_Task, 'duration_secs': 0.421822} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.679939] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Reconfigured VM instance instance-00000076 to attach disk [datastore2] 7b82801d-7a0f-4baf-ad2e-566174d3eb35/7b82801d-7a0f-4baf-ad2e-566174d3eb35.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1244.680568] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ace69e77-c4b2-400c-8cef-2195e3c77c82 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.686640] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1244.686640] env[62814]: value = "task-4294808" [ 1244.686640] env[62814]: _type = "Task" [ 1244.686640] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.695350] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294808, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.059805] env[62814]: DEBUG oslo_concurrency.lockutils [None req-90065ee8-1047-4dac-b3f5-590b232e0a3e tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "b1d88997-e52f-41bd-b1b4-dd096d20d60a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.685s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1245.102040] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1245.102272] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1245.102437] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1245.102623] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1245.102769] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1245.102913] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1245.103132] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1245.103310] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1245.103450] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1245.103609] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1245.103784] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1245.109162] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-faa16462-c04b-474a-beff-b922da194392 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.128193] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1245.128193] env[62814]: value = "task-4294809" [ 1245.128193] env[62814]: _type = "Task" [ 1245.128193] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.138530] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294809, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.156892] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cbf40bd-0b23-4eb5-b339-2c60aebfdd2c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.165155] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-031a5dbc-e4e3-45ac-aa3d-46d26b802501 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.198978] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22fd5ac0-e3fc-47b4-9cda-6b7927fc3ef5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.206332] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294808, 'name': Rename_Task, 'duration_secs': 0.277866} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.208724] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1245.208809] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fcb54700-425c-478f-a92a-913e63f63d98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.211106] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-992a6a22-d9da-4301-8014-051c387e4178 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.227351] env[62814]: DEBUG nova.compute.provider_tree [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1245.227397] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1245.227397] env[62814]: value = "task-4294810" [ 1245.227397] env[62814]: _type = "Task" [ 1245.227397] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.235336] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294810, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.588661] env[62814]: DEBUG oslo_concurrency.lockutils [None req-19f535fc-1739-4423-bd34-5f3c4c50813c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.286s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1245.637874] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294809, 'name': ReconfigVM_Task, 'duration_secs': 0.233257} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.638241] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance 'ae9180b8-6caa-44ef-93a8-eb3a7681d224' progress to 33 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1245.728935] env[62814]: DEBUG nova.scheduler.client.report [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1245.741025] env[62814]: DEBUG oslo_vmware.api [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294810, 'name': PowerOnVM_Task, 'duration_secs': 0.442582} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.742591] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1245.742591] env[62814]: INFO nova.compute.manager [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Took 7.54 seconds to spawn the instance on the hypervisor. [ 1245.742591] env[62814]: DEBUG nova.compute.manager [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1245.742913] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67f2e392-9cf8-48e3-abf3-fc477f42c44f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.905126] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "d043b12d-a316-46ab-a30f-abec0a7963a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1245.905369] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1245.997934] env[62814]: DEBUG oslo_concurrency.lockutils [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1245.998193] env[62814]: DEBUG oslo_concurrency.lockutils [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1246.144708] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1246.145106] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1246.145179] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1246.145367] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1246.145545] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1246.145699] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1246.145905] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1246.146076] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1246.146249] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1246.146412] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1246.146868] env[62814]: DEBUG nova.virt.hardware [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1246.151969] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Reconfiguring VM instance instance-00000075 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1246.152511] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-791e0e52-de8a-4080-a0e9-36234dd4b9c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.171420] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1246.171420] env[62814]: value = "task-4294811" [ 1246.171420] env[62814]: _type = "Task" [ 1246.171420] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.179281] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294811, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.234130] env[62814]: DEBUG oslo_concurrency.lockutils [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.706s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1246.262992] env[62814]: INFO nova.compute.manager [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Took 12.79 seconds to build instance. [ 1246.269902] env[62814]: INFO nova.scheduler.client.report [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleted allocations for instance 606e31eb-2349-427f-9c9b-ed9dc5b385f5 [ 1246.408220] env[62814]: DEBUG nova.compute.manager [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1246.501064] env[62814]: DEBUG nova.compute.utils [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1246.612691] env[62814]: DEBUG oslo_concurrency.lockutils [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "1f13051e-e51b-4981-9445-d5420c3c1818" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1246.613190] env[62814]: DEBUG oslo_concurrency.lockutils [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1246.613424] env[62814]: DEBUG oslo_concurrency.lockutils [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1246.613620] env[62814]: DEBUG oslo_concurrency.lockutils [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1246.613849] env[62814]: DEBUG oslo_concurrency.lockutils [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1246.618738] env[62814]: INFO nova.compute.manager [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Terminating instance [ 1246.681318] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294811, 'name': ReconfigVM_Task, 'duration_secs': 0.180292} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.681632] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Reconfigured VM instance instance-00000075 to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1246.682925] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-879dbc73-99b2-479b-961a-ef3500156088 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.705984] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Reconfiguring VM instance instance-00000075 to attach disk [datastore2] ae9180b8-6caa-44ef-93a8-eb3a7681d224/ae9180b8-6caa-44ef-93a8-eb3a7681d224.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1246.706276] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-10cc5726-9bc7-474d-9453-3106eeee5093 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.723515] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1246.723515] env[62814]: value = "task-4294812" [ 1246.723515] env[62814]: _type = "Task" [ 1246.723515] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.731211] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294812, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.765480] env[62814]: DEBUG oslo_concurrency.lockutils [None req-e6ab87d1-d15a-454d-8188-f6ef438c86f8 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.314s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1246.776678] env[62814]: DEBUG oslo_concurrency.lockutils [None req-09ef4f37-a0ad-4e52-9aea-a9607f5361af tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "606e31eb-2349-427f-9c9b-ed9dc5b385f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.738s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1246.933369] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1246.933650] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1246.935303] env[62814]: INFO nova.compute.claims [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1247.006884] env[62814]: DEBUG oslo_concurrency.lockutils [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1247.123056] env[62814]: DEBUG nova.compute.manager [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1247.123288] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1247.124223] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709e79b3-efa5-49ee-a58e-124f45f2a16d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.135544] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1247.135787] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa9f075a-0dda-42bc-a0e2-560477970e9d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.146367] env[62814]: DEBUG oslo_vmware.api [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1247.146367] env[62814]: value = "task-4294813" [ 1247.146367] env[62814]: _type = "Task" [ 1247.146367] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.156521] env[62814]: DEBUG oslo_vmware.api [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294813, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.232756] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294812, 'name': ReconfigVM_Task, 'duration_secs': 0.313971} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.233065] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Reconfigured VM instance instance-00000075 to attach disk [datastore2] ae9180b8-6caa-44ef-93a8-eb3a7681d224/ae9180b8-6caa-44ef-93a8-eb3a7681d224.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1247.233392] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance 'ae9180b8-6caa-44ef-93a8-eb3a7681d224' progress to 50 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1247.478704] env[62814]: DEBUG nova.compute.manager [req-2dd5eb9b-8f50-4407-8bb9-1f2291318514 req-99affdee-992c-4490-85b9-9769c8c8d3cd service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received event network-changed-9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1247.478907] env[62814]: DEBUG nova.compute.manager [req-2dd5eb9b-8f50-4407-8bb9-1f2291318514 req-99affdee-992c-4490-85b9-9769c8c8d3cd service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing instance network info cache due to event network-changed-9edabfa8-9760-4ef4-adfb-afc6e2200b50. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1247.479624] env[62814]: DEBUG oslo_concurrency.lockutils [req-2dd5eb9b-8f50-4407-8bb9-1f2291318514 req-99affdee-992c-4490-85b9-9769c8c8d3cd service nova] Acquiring lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1247.479624] env[62814]: DEBUG oslo_concurrency.lockutils [req-2dd5eb9b-8f50-4407-8bb9-1f2291318514 req-99affdee-992c-4490-85b9-9769c8c8d3cd service nova] Acquired lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1247.479624] env[62814]: DEBUG nova.network.neutron [req-2dd5eb9b-8f50-4407-8bb9-1f2291318514 req-99affdee-992c-4490-85b9-9769c8c8d3cd service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing network info cache for port 9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1247.660936] env[62814]: DEBUG oslo_vmware.api [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294813, 'name': PowerOffVM_Task, 'duration_secs': 0.322483} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.661742] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1247.661742] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1247.661872] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8719e519-9e30-4040-947b-21ed8adb9762 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.741782] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bfe0b48-0ecd-49dc-a439-bc251b2e25f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.744278] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1247.744514] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1247.744706] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleting the datastore file [datastore2] 1f13051e-e51b-4981-9445-d5420c3c1818 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1247.744955] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6fc6a68c-7787-4a92-8a08-0ab82b2c42b6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.764313] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356e045e-7795-4e05-9a4b-ba75a7d5e8a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.767038] env[62814]: DEBUG oslo_vmware.api [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1247.767038] env[62814]: value = "task-4294815" [ 1247.767038] env[62814]: _type = "Task" [ 1247.767038] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.783287] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance 'ae9180b8-6caa-44ef-93a8-eb3a7681d224' progress to 67 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1247.791304] env[62814]: DEBUG oslo_vmware.api [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294815, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.927957] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "add41cff-326f-46f4-b768-1f08ace2b31f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1247.928169] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "add41cff-326f-46f4-b768-1f08ace2b31f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1248.058014] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90489777-2f0f-4d0a-914c-63167a55cce0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.065783] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28298562-9f3a-4c4d-af5b-2148b09b9df1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.096769] env[62814]: DEBUG oslo_concurrency.lockutils [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1248.097013] env[62814]: DEBUG oslo_concurrency.lockutils [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1248.097242] env[62814]: INFO nova.compute.manager [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Attaching volume 0c2dd27e-67d6-437e-bebb-4a9faa22c7d2 to /dev/sdb [ 1248.101882] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d6bd0a-624e-44a4-b596-ee03160764e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.111372] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f146f808-f8ad-443c-898d-95972d16424b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.124396] env[62814]: DEBUG nova.compute.provider_tree [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1248.140799] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48635b02-c7b2-4b6d-84d1-c9d01ee81ca6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.147578] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51fe4e0d-7b27-4db4-8a79-8f04074dd38e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.161435] env[62814]: DEBUG nova.virt.block_device [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Updating existing volume attachment record: f92da50a-d4a0-4001-8734-5c2de077094f {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1248.203582] env[62814]: DEBUG nova.network.neutron [req-2dd5eb9b-8f50-4407-8bb9-1f2291318514 req-99affdee-992c-4490-85b9-9769c8c8d3cd service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updated VIF entry in instance network info cache for port 9edabfa8-9760-4ef4-adfb-afc6e2200b50. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1248.203904] env[62814]: DEBUG nova.network.neutron [req-2dd5eb9b-8f50-4407-8bb9-1f2291318514 req-99affdee-992c-4490-85b9-9769c8c8d3cd service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [{"id": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "address": "fa:16:3e:ff:ce:8d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9edabfa8-97", "ovs_interfaceid": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1248.277142] env[62814]: DEBUG oslo_vmware.api [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294815, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.388986} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.277545] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1248.277666] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1248.277873] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1248.278099] env[62814]: INFO nova.compute.manager [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1248.278352] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1248.278543] env[62814]: DEBUG nova.compute.manager [-] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1248.278638] env[62814]: DEBUG nova.network.neutron [-] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1248.431133] env[62814]: DEBUG nova.compute.manager [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1248.505418] env[62814]: DEBUG nova.network.neutron [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Port 3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8 binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1248.628811] env[62814]: DEBUG nova.scheduler.client.report [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1248.706633] env[62814]: DEBUG oslo_concurrency.lockutils [req-2dd5eb9b-8f50-4407-8bb9-1f2291318514 req-99affdee-992c-4490-85b9-9769c8c8d3cd service nova] Releasing lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1248.873950] env[62814]: DEBUG nova.compute.manager [req-d43547f6-8239-4b1a-87ea-97864255e779 req-56caf114-1306-4f24-9847-190199c433f0 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Received event network-vif-deleted-a7104ea3-596d-4d10-952c-ac861087e580 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1248.874164] env[62814]: INFO nova.compute.manager [req-d43547f6-8239-4b1a-87ea-97864255e779 req-56caf114-1306-4f24-9847-190199c433f0 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Neutron deleted interface a7104ea3-596d-4d10-952c-ac861087e580; detaching it from the instance and deleting it from the info cache [ 1248.874341] env[62814]: DEBUG nova.network.neutron [req-d43547f6-8239-4b1a-87ea-97864255e779 req-56caf114-1306-4f24-9847-190199c433f0 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1248.957526] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1249.139580] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1249.140102] env[62814]: DEBUG nova.compute.manager [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1249.142793] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.185s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1249.144259] env[62814]: INFO nova.compute.claims [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1249.349671] env[62814]: DEBUG nova.network.neutron [-] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.378140] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-98f5a2e3-47a9-4225-8605-6a7449b9e39d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.387795] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ec6b21b-b44b-4392-82b4-126fe30d4a14 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.415940] env[62814]: DEBUG nova.compute.manager [req-d43547f6-8239-4b1a-87ea-97864255e779 req-56caf114-1306-4f24-9847-190199c433f0 service nova] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Detach interface failed, port_id=a7104ea3-596d-4d10-952c-ac861087e580, reason: Instance 1f13051e-e51b-4981-9445-d5420c3c1818 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1249.509475] env[62814]: DEBUG nova.compute.manager [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Received event network-changed-90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1249.509475] env[62814]: DEBUG nova.compute.manager [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Refreshing instance network info cache due to event network-changed-90c053d0-8e56-42b1-89da-b7acd45ac0fa. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1249.509475] env[62814]: DEBUG oslo_concurrency.lockutils [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] Acquiring lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1249.509475] env[62814]: DEBUG oslo_concurrency.lockutils [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] Acquired lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1249.509475] env[62814]: DEBUG nova.network.neutron [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Refreshing network info cache for port 90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1249.525905] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1249.526393] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1249.526733] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1249.652187] env[62814]: DEBUG nova.compute.utils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1249.654668] env[62814]: DEBUG nova.compute.manager [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1249.655109] env[62814]: DEBUG nova.network.neutron [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1249.707624] env[62814]: DEBUG nova.policy [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9df1642a53c940b2bfacc82e1ac975d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '446e63a428c34d38ad07d01f0c90b23c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1249.852137] env[62814]: INFO nova.compute.manager [-] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Took 1.57 seconds to deallocate network for instance. [ 1250.008937] env[62814]: DEBUG nova.network.neutron [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Successfully created port: d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1250.155630] env[62814]: DEBUG nova.compute.manager [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1250.203445] env[62814]: DEBUG nova.network.neutron [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updated VIF entry in instance network info cache for port 90c053d0-8e56-42b1-89da-b7acd45ac0fa. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1250.203800] env[62814]: DEBUG nova.network.neutron [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updating instance_info_cache with network_info: [{"id": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "address": "fa:16:3e:e7:24:4d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c053d0-8e", "ovs_interfaceid": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1250.273534] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9245b6-6179-472b-ab36-ecddc425d750 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.281316] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbee3e6-c4f5-49f2-822a-425392223ef1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.311514] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2e27a7-68d4-4ffa-89c0-44f5924a1cf3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.318347] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2409d5d9-ba74-4c76-8e51-f4b59694942f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.331344] env[62814]: DEBUG nova.compute.provider_tree [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1250.358098] env[62814]: DEBUG oslo_concurrency.lockutils [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1250.565510] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.566826] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1250.566826] env[62814]: DEBUG nova.network.neutron [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1250.706043] env[62814]: DEBUG oslo_concurrency.lockutils [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] Releasing lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1250.706313] env[62814]: DEBUG nova.compute.manager [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Received event network-changed-90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1250.706485] env[62814]: DEBUG nova.compute.manager [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Refreshing instance network info cache due to event network-changed-90c053d0-8e56-42b1-89da-b7acd45ac0fa. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1250.706692] env[62814]: DEBUG oslo_concurrency.lockutils [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] Acquiring lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1250.706833] env[62814]: DEBUG oslo_concurrency.lockutils [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] Acquired lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1250.706995] env[62814]: DEBUG nova.network.neutron [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Refreshing network info cache for port 90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1250.834858] env[62814]: DEBUG nova.scheduler.client.report [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1251.169181] env[62814]: DEBUG nova.compute.manager [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1251.196958] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1251.197226] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1251.197446] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1251.197688] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1251.197849] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1251.198008] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1251.198235] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1251.198396] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1251.198566] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1251.198740] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1251.198916] env[62814]: DEBUG nova.virt.hardware [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1251.199874] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf78e207-7947-495c-b816-448105cefb2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.212324] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a58b4ae-be0c-4ae7-8725-9b1a99de5ff0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.313605] env[62814]: DEBUG nova.network.neutron [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance_info_cache with network_info: [{"id": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "address": "fa:16:3e:94:da:3d", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d17bc61-a2", "ovs_interfaceid": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1251.340373] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.197s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1251.340880] env[62814]: DEBUG nova.compute.manager [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1251.343447] env[62814]: DEBUG oslo_concurrency.lockutils [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.985s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1251.343694] env[62814]: DEBUG nova.objects.instance [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'resources' on Instance uuid 1f13051e-e51b-4981-9445-d5420c3c1818 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1251.381184] env[62814]: DEBUG nova.compute.manager [req-62792b26-a8c8-412a-b247-ab832d65f25c req-5671a73e-f2ed-472e-aa37-96b796d9b053 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received event network-vif-plugged-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1251.381459] env[62814]: DEBUG oslo_concurrency.lockutils [req-62792b26-a8c8-412a-b247-ab832d65f25c req-5671a73e-f2ed-472e-aa37-96b796d9b053 service nova] Acquiring lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1251.381606] env[62814]: DEBUG oslo_concurrency.lockutils [req-62792b26-a8c8-412a-b247-ab832d65f25c req-5671a73e-f2ed-472e-aa37-96b796d9b053 service nova] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1251.381775] env[62814]: DEBUG oslo_concurrency.lockutils [req-62792b26-a8c8-412a-b247-ab832d65f25c req-5671a73e-f2ed-472e-aa37-96b796d9b053 service nova] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1251.381934] env[62814]: DEBUG nova.compute.manager [req-62792b26-a8c8-412a-b247-ab832d65f25c req-5671a73e-f2ed-472e-aa37-96b796d9b053 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] No waiting events found dispatching network-vif-plugged-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1251.382111] env[62814]: WARNING nova.compute.manager [req-62792b26-a8c8-412a-b247-ab832d65f25c req-5671a73e-f2ed-472e-aa37-96b796d9b053 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received unexpected event network-vif-plugged-d584963d-9840-4607-b75d-b6a46b94cea3 for instance with vm_state building and task_state spawning. [ 1251.437125] env[62814]: DEBUG nova.network.neutron [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updated VIF entry in instance network info cache for port 90c053d0-8e56-42b1-89da-b7acd45ac0fa. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1251.437485] env[62814]: DEBUG nova.network.neutron [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updating instance_info_cache with network_info: [{"id": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "address": "fa:16:3e:e7:24:4d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c053d0-8e", "ovs_interfaceid": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1251.456759] env[62814]: DEBUG nova.network.neutron [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Successfully updated port: d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1251.816040] env[62814]: DEBUG oslo_concurrency.lockutils [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1251.851536] env[62814]: DEBUG nova.compute.utils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1251.853881] env[62814]: DEBUG nova.compute.manager [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1251.853881] env[62814]: DEBUG nova.network.neutron [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1251.900899] env[62814]: DEBUG nova.policy [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '51348be2493643fb9e5b44f27e5a7a94', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '513135674ee446d19fa8c667a47138a5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1251.939761] env[62814]: DEBUG oslo_concurrency.lockutils [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] Releasing lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1251.940042] env[62814]: DEBUG nova.compute.manager [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received event network-changed-9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1251.940219] env[62814]: DEBUG nova.compute.manager [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing instance network info cache due to event network-changed-9edabfa8-9760-4ef4-adfb-afc6e2200b50. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1251.940423] env[62814]: DEBUG oslo_concurrency.lockutils [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] Acquiring lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.940562] env[62814]: DEBUG oslo_concurrency.lockutils [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] Acquired lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1251.940722] env[62814]: DEBUG nova.network.neutron [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing network info cache for port 9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1251.959690] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.959833] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1251.959977] env[62814]: DEBUG nova.network.neutron [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1251.970059] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65cd6478-63ca-4d68-919a-7221ae889cba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.978554] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66cfae99-1633-4bde-a3f8-9c68ff3f9e5e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.008568] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e3bd4c-d38e-4c9d-83cf-8bcf66153c72 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.016130] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494446d3-f982-4acb-bd02-15cb657f6fb9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.029602] env[62814]: DEBUG nova.compute.provider_tree [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1252.209377] env[62814]: DEBUG nova.network.neutron [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Successfully created port: 14c4e34d-9d53-41a8-ade9-86a6fa15cab5 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1252.337983] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8de4f766-9ad3-485f-a861-c963ab5265f3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.358416] env[62814]: DEBUG nova.compute.manager [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1252.364301] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630ce7c1-778e-4ed8-a468-7fbbab0c92a8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.370735] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance 'ae9180b8-6caa-44ef-93a8-eb3a7681d224' progress to 83 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1252.501289] env[62814]: DEBUG nova.network.neutron [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1252.533041] env[62814]: DEBUG nova.scheduler.client.report [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1252.711113] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1252.711113] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845869', 'volume_id': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'name': 'volume-0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '0fdb57c1-7a2a-455d-acb8-9f342ef1dbac', 'attached_at': '', 'detached_at': '', 'volume_id': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'serial': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1252.711113] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6b8310-1e68-481e-8337-2f57f8ba6372 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.730048] env[62814]: DEBUG nova.network.neutron [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updating instance_info_cache with network_info: [{"id": "d584963d-9840-4607-b75d-b6a46b94cea3", "address": "fa:16:3e:ca:e5:53", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd584963d-98", "ovs_interfaceid": "d584963d-9840-4607-b75d-b6a46b94cea3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1252.731300] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6b9673-a343-45a4-adf8-1803d50de09a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.757084] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] volume-0c2dd27e-67d6-437e-bebb-4a9faa22c7d2/volume-0c2dd27e-67d6-437e-bebb-4a9faa22c7d2.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1252.758292] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e85cf760-17d6-4c75-9e75-5bf07319fbb4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.771382] env[62814]: DEBUG nova.network.neutron [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updated VIF entry in instance network info cache for port 9edabfa8-9760-4ef4-adfb-afc6e2200b50. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1252.771773] env[62814]: DEBUG nova.network.neutron [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [{"id": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "address": "fa:16:3e:ff:ce:8d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9edabfa8-97", "ovs_interfaceid": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1252.779161] env[62814]: DEBUG oslo_vmware.api [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1252.779161] env[62814]: value = "task-4294818" [ 1252.779161] env[62814]: _type = "Task" [ 1252.779161] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.788364] env[62814]: DEBUG oslo_vmware.api [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294818, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.876737] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1252.876991] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92e80b96-14a0-4989-a9e0-acbc29693093 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1252.883975] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1252.883975] env[62814]: value = "task-4294819" [ 1252.883975] env[62814]: _type = "Task" [ 1252.883975] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1252.892052] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294819, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.038047] env[62814]: DEBUG oslo_concurrency.lockutils [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.694s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.063677] env[62814]: INFO nova.scheduler.client.report [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleted allocations for instance 1f13051e-e51b-4981-9445-d5420c3c1818 [ 1253.235140] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1253.235374] env[62814]: DEBUG nova.compute.manager [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Instance network_info: |[{"id": "d584963d-9840-4607-b75d-b6a46b94cea3", "address": "fa:16:3e:ca:e5:53", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd584963d-98", "ovs_interfaceid": "d584963d-9840-4607-b75d-b6a46b94cea3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1253.235863] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:e5:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b83383f-ed7a-4efd-aef7-aa8c15649d07', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd584963d-9840-4607-b75d-b6a46b94cea3', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1253.244021] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1253.244262] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1253.244550] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-10fc7bc1-d96b-4d03-81e5-ffa899dfc09a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.267244] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1253.267244] env[62814]: value = "task-4294820" [ 1253.267244] env[62814]: _type = "Task" [ 1253.267244] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.275293] env[62814]: DEBUG oslo_concurrency.lockutils [req-def3f8e6-069b-4eb0-8414-9a051e77ea7b req-f7abcd29-0d28-41f5-a747-51f6fe6d8df7 service nova] Releasing lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1253.275649] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294820, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.287250] env[62814]: DEBUG oslo_vmware.api [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294818, 'name': ReconfigVM_Task, 'duration_secs': 0.350782} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.287506] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Reconfigured VM instance instance-00000071 to attach disk [datastore1] volume-0c2dd27e-67d6-437e-bebb-4a9faa22c7d2/volume-0c2dd27e-67d6-437e-bebb-4a9faa22c7d2.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1253.292044] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2e55f7c-980c-47b9-a016-01b110df391b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.307499] env[62814]: DEBUG oslo_vmware.api [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1253.307499] env[62814]: value = "task-4294821" [ 1253.307499] env[62814]: _type = "Task" [ 1253.307499] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.317141] env[62814]: DEBUG oslo_vmware.api [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294821, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.374315] env[62814]: DEBUG nova.compute.manager [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1253.394480] env[62814]: DEBUG oslo_vmware.api [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294819, 'name': PowerOnVM_Task, 'duration_secs': 0.372049} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.395016] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1253.395184] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-892235f7-621f-4cd6-9d62-60160404c378 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance 'ae9180b8-6caa-44ef-93a8-eb3a7681d224' progress to 100 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1253.408396] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1253.408675] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1253.408862] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1253.409070] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1253.409244] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1253.409478] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1253.409721] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1253.409893] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1253.410073] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1253.410240] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1253.410412] env[62814]: DEBUG nova.virt.hardware [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1253.411287] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f785a9-b7b3-4d84-845b-bcebbd2ece14 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.416123] env[62814]: DEBUG nova.compute.manager [req-f74626ee-eea8-4bf8-b9db-f3bfb79bc3f6 req-c58f1df1-50d3-4958-adcb-b9446e774e33 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received event network-changed-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1253.416781] env[62814]: DEBUG nova.compute.manager [req-f74626ee-eea8-4bf8-b9db-f3bfb79bc3f6 req-c58f1df1-50d3-4958-adcb-b9446e774e33 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Refreshing instance network info cache due to event network-changed-d584963d-9840-4607-b75d-b6a46b94cea3. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1253.417048] env[62814]: DEBUG oslo_concurrency.lockutils [req-f74626ee-eea8-4bf8-b9db-f3bfb79bc3f6 req-c58f1df1-50d3-4958-adcb-b9446e774e33 service nova] Acquiring lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1253.417214] env[62814]: DEBUG oslo_concurrency.lockutils [req-f74626ee-eea8-4bf8-b9db-f3bfb79bc3f6 req-c58f1df1-50d3-4958-adcb-b9446e774e33 service nova] Acquired lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1253.417381] env[62814]: DEBUG nova.network.neutron [req-f74626ee-eea8-4bf8-b9db-f3bfb79bc3f6 req-c58f1df1-50d3-4958-adcb-b9446e774e33 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Refreshing network info cache for port d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1253.423897] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2defe373-a6f5-4514-9a8c-70126415a231 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.575319] env[62814]: DEBUG oslo_concurrency.lockutils [None req-49c727da-8c1f-44e7-9006-6ff88a707e72 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "1f13051e-e51b-4981-9445-d5420c3c1818" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.962s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1253.718655] env[62814]: DEBUG nova.network.neutron [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Successfully updated port: 14c4e34d-9d53-41a8-ade9-86a6fa15cab5 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1253.778052] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294820, 'name': CreateVM_Task, 'duration_secs': 0.483217} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.778052] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1253.778492] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1253.778659] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1253.778978] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1253.779254] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8877dd05-eb0f-4fcd-b5ed-3cce93559d17 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.783754] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1253.783754] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525b5bb4-0f35-7edf-1d72-a54364d68971" [ 1253.783754] env[62814]: _type = "Task" [ 1253.783754] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1253.791374] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525b5bb4-0f35-7edf-1d72-a54364d68971, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1253.815842] env[62814]: DEBUG oslo_vmware.api [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294821, 'name': ReconfigVM_Task, 'duration_secs': 0.151572} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1253.816129] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845869', 'volume_id': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'name': 'volume-0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '0fdb57c1-7a2a-455d-acb8-9f342ef1dbac', 'attached_at': '', 'detached_at': '', 'volume_id': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'serial': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1254.176044] env[62814]: DEBUG nova.network.neutron [req-f74626ee-eea8-4bf8-b9db-f3bfb79bc3f6 req-c58f1df1-50d3-4958-adcb-b9446e774e33 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updated VIF entry in instance network info cache for port d584963d-9840-4607-b75d-b6a46b94cea3. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1254.176432] env[62814]: DEBUG nova.network.neutron [req-f74626ee-eea8-4bf8-b9db-f3bfb79bc3f6 req-c58f1df1-50d3-4958-adcb-b9446e774e33 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updating instance_info_cache with network_info: [{"id": "d584963d-9840-4607-b75d-b6a46b94cea3", "address": "fa:16:3e:ca:e5:53", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd584963d-98", "ovs_interfaceid": "d584963d-9840-4607-b75d-b6a46b94cea3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1254.222482] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.222848] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1254.222848] env[62814]: DEBUG nova.network.neutron [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1254.295647] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]525b5bb4-0f35-7edf-1d72-a54364d68971, 'name': SearchDatastore_Task, 'duration_secs': 0.010676} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.296085] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1254.296395] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1254.296718] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.296969] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1254.297263] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1254.297870] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d856f4b-de15-4dda-84c0-86a9f7133e0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.307169] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1254.307364] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1254.308101] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91c4325b-8c76-4292-8cc8-eef22b607349 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.313686] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1254.313686] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521fb8f7-8b8a-2ab3-1884-23f9fe0c388c" [ 1254.313686] env[62814]: _type = "Task" [ 1254.313686] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.323727] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521fb8f7-8b8a-2ab3-1884-23f9fe0c388c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.679674] env[62814]: DEBUG oslo_concurrency.lockutils [req-f74626ee-eea8-4bf8-b9db-f3bfb79bc3f6 req-c58f1df1-50d3-4958-adcb-b9446e774e33 service nova] Releasing lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1254.772523] env[62814]: DEBUG nova.network.neutron [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1254.824151] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]521fb8f7-8b8a-2ab3-1884-23f9fe0c388c, 'name': SearchDatastore_Task, 'duration_secs': 0.008793} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1254.827147] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-018d9230-8b09-4fb4-bc3f-8d701205ef71 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.834394] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1254.834394] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5239866d-fef9-28ee-f3ca-8a3c800d4225" [ 1254.834394] env[62814]: _type = "Task" [ 1254.834394] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1254.842035] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5239866d-fef9-28ee-f3ca-8a3c800d4225, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1254.860016] env[62814]: DEBUG nova.objects.instance [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lazy-loading 'flavor' on Instance uuid 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1254.928582] env[62814]: DEBUG nova.network.neutron [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Updating instance_info_cache with network_info: [{"id": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "address": "fa:16:3e:c3:d9:f1", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14c4e34d-9d", "ovs_interfaceid": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.350414] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5239866d-fef9-28ee-f3ca-8a3c800d4225, 'name': SearchDatastore_Task, 'duration_secs': 0.01105} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.351391] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1255.351804] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d043b12d-a316-46ab-a30f-abec0a7963a6/d043b12d-a316-46ab-a30f-abec0a7963a6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1255.352113] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc197b4c-d5ba-4389-9226-d27bb45d5f69 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.360162] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1255.360162] env[62814]: value = "task-4294822" [ 1255.360162] env[62814]: _type = "Task" [ 1255.360162] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.369014] env[62814]: DEBUG oslo_concurrency.lockutils [None req-43784f67-1afd-4fcf-9cda-00c2d8001fcc tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.272s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1255.373887] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294822, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.431307] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1255.431658] env[62814]: DEBUG nova.compute.manager [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Instance network_info: |[{"id": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "address": "fa:16:3e:c3:d9:f1", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14c4e34d-9d", "ovs_interfaceid": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1255.432163] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:d9:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5b21ab10-d886-4453-9472-9e11fb3c450d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '14c4e34d-9d53-41a8-ade9-86a6fa15cab5', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1255.440210] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1255.440458] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1255.440710] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-785b801f-538b-4328-91d9-601543f1f9a3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.462461] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1255.462461] env[62814]: value = "task-4294823" [ 1255.462461] env[62814]: _type = "Task" [ 1255.462461] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.470529] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294823, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.595452] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1255.595826] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1255.807618] env[62814]: DEBUG nova.compute.manager [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Received event network-vif-plugged-14c4e34d-9d53-41a8-ade9-86a6fa15cab5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1255.807993] env[62814]: DEBUG oslo_concurrency.lockutils [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] Acquiring lock "add41cff-326f-46f4-b768-1f08ace2b31f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1255.807993] env[62814]: DEBUG oslo_concurrency.lockutils [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] Lock "add41cff-326f-46f4-b768-1f08ace2b31f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1255.808148] env[62814]: DEBUG oslo_concurrency.lockutils [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] Lock "add41cff-326f-46f4-b768-1f08ace2b31f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1255.808325] env[62814]: DEBUG nova.compute.manager [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] No waiting events found dispatching network-vif-plugged-14c4e34d-9d53-41a8-ade9-86a6fa15cab5 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1255.808488] env[62814]: WARNING nova.compute.manager [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Received unexpected event network-vif-plugged-14c4e34d-9d53-41a8-ade9-86a6fa15cab5 for instance with vm_state building and task_state spawning. [ 1255.808647] env[62814]: DEBUG nova.compute.manager [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Received event network-changed-14c4e34d-9d53-41a8-ade9-86a6fa15cab5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1255.808965] env[62814]: DEBUG nova.compute.manager [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Refreshing instance network info cache due to event network-changed-14c4e34d-9d53-41a8-ade9-86a6fa15cab5. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1255.809093] env[62814]: DEBUG oslo_concurrency.lockutils [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] Acquiring lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.809239] env[62814]: DEBUG oslo_concurrency.lockutils [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] Acquired lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1255.809395] env[62814]: DEBUG nova.network.neutron [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Refreshing network info cache for port 14c4e34d-9d53-41a8-ade9-86a6fa15cab5 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1255.840031] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1255.840295] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1255.840483] env[62814]: DEBUG nova.compute.manager [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Going to confirm migration 9 {{(pid=62814) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 1255.869723] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294822, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476324} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.870043] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] d043b12d-a316-46ab-a30f-abec0a7963a6/d043b12d-a316-46ab-a30f-abec0a7963a6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1255.870282] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1255.870791] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6a634e7c-8632-49ed-ac6b-f7dcd15ab630 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.876785] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1255.876785] env[62814]: value = "task-4294824" [ 1255.876785] env[62814]: _type = "Task" [ 1255.876785] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.885759] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294824, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.972359] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294823, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.099616] env[62814]: INFO nova.compute.manager [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Detaching volume 0c2dd27e-67d6-437e-bebb-4a9faa22c7d2 [ 1256.136630] env[62814]: INFO nova.virt.block_device [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Attempting to driver detach volume 0c2dd27e-67d6-437e-bebb-4a9faa22c7d2 from mountpoint /dev/sdb [ 1256.136630] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1256.136774] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845869', 'volume_id': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'name': 'volume-0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '0fdb57c1-7a2a-455d-acb8-9f342ef1dbac', 'attached_at': '', 'detached_at': '', 'volume_id': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'serial': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1256.137658] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc5ddde-9a53-4665-8f70-48ed2e2c6bc3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.159319] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd2f1ad-dce1-4320-aeb0-54e8015cb2ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.166292] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86abc353-bf7f-4972-a602-c1eed3d61ba7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.188089] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5baab7-ad6d-4940-8390-7db041d5d303 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.203027] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] The volume has not been displaced from its original location: [datastore1] volume-0c2dd27e-67d6-437e-bebb-4a9faa22c7d2/volume-0c2dd27e-67d6-437e-bebb-4a9faa22c7d2.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1256.208218] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Reconfiguring VM instance instance-00000071 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1256.208531] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f990f13-05bd-4345-b4d9-12edb0135c60 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.226468] env[62814]: DEBUG oslo_vmware.api [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1256.226468] env[62814]: value = "task-4294825" [ 1256.226468] env[62814]: _type = "Task" [ 1256.226468] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.233947] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1256.234192] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1256.238822] env[62814]: DEBUG oslo_vmware.api [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294825, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.375728] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.375951] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquired lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1256.376166] env[62814]: DEBUG nova.network.neutron [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1256.376358] env[62814]: DEBUG nova.objects.instance [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lazy-loading 'info_cache' on Instance uuid ae9180b8-6caa-44ef-93a8-eb3a7681d224 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1256.389112] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294824, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068907} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.389922] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1256.390807] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cbda052-e45f-44c9-a138-6215d6a6e9ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.413140] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] d043b12d-a316-46ab-a30f-abec0a7963a6/d043b12d-a316-46ab-a30f-abec0a7963a6.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1256.415656] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0306c28b-0330-4931-9cd8-8a0908db7dde {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.434053] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1256.434053] env[62814]: value = "task-4294826" [ 1256.434053] env[62814]: _type = "Task" [ 1256.434053] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.441525] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294826, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.472688] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294823, 'name': CreateVM_Task, 'duration_secs': 0.546324} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.472893] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1256.473538] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.473701] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1256.474052] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1256.474256] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00ebe3c3-6483-430a-9cbe-b20691ed3bc4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.481172] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1256.481172] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e9b196-c761-a2ec-f6ae-ffd197c8d521" [ 1256.481172] env[62814]: _type = "Task" [ 1256.481172] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.488552] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e9b196-c761-a2ec-f6ae-ffd197c8d521, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.536799] env[62814]: DEBUG nova.network.neutron [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Updated VIF entry in instance network info cache for port 14c4e34d-9d53-41a8-ade9-86a6fa15cab5. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1256.537260] env[62814]: DEBUG nova.network.neutron [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Updating instance_info_cache with network_info: [{"id": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "address": "fa:16:3e:c3:d9:f1", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14c4e34d-9d", "ovs_interfaceid": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1256.738174] env[62814]: DEBUG oslo_vmware.api [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294825, 'name': ReconfigVM_Task, 'duration_secs': 0.199948} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.738478] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Reconfigured VM instance instance-00000071 to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1256.743202] env[62814]: DEBUG nova.compute.manager [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1256.745853] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32b354bb-3846-4803-a199-c4e55f31869b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.760520] env[62814]: DEBUG oslo_vmware.api [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1256.760520] env[62814]: value = "task-4294827" [ 1256.760520] env[62814]: _type = "Task" [ 1256.760520] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.768471] env[62814]: DEBUG oslo_vmware.api [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294827, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.943623] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294826, 'name': ReconfigVM_Task, 'duration_secs': 0.256911} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.944045] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Reconfigured VM instance instance-00000077 to attach disk [datastore2] d043b12d-a316-46ab-a30f-abec0a7963a6/d043b12d-a316-46ab-a30f-abec0a7963a6.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1256.944280] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c5eaca9-6456-47b6-a8d8-da2efa5edad2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.949904] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1256.949904] env[62814]: value = "task-4294828" [ 1256.949904] env[62814]: _type = "Task" [ 1256.949904] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.956978] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294828, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.990346] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e9b196-c761-a2ec-f6ae-ffd197c8d521, 'name': SearchDatastore_Task, 'duration_secs': 0.009379} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.990657] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1256.990884] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1256.991128] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.991277] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1256.991458] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1256.991725] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c1d1f2f8-aa78-40d1-9138-6fc3be50cfc9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.999687] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1256.999912] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1257.000593] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94975b23-e482-4714-a6f2-eca36f040ea2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.005315] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1257.005315] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52634ffe-027e-2fd7-cbe2-d9b3483fe7bd" [ 1257.005315] env[62814]: _type = "Task" [ 1257.005315] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.012443] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52634ffe-027e-2fd7-cbe2-d9b3483fe7bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.039482] env[62814]: DEBUG oslo_concurrency.lockutils [req-cc1050dc-dc44-440b-ae18-25301578cbea req-a9a76e0d-ae40-4622-a492-840e322aa9a0 service nova] Releasing lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1257.271886] env[62814]: DEBUG oslo_vmware.api [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294827, 'name': ReconfigVM_Task, 'duration_secs': 0.13918} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.272231] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845869', 'volume_id': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'name': 'volume-0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '0fdb57c1-7a2a-455d-acb8-9f342ef1dbac', 'attached_at': '', 'detached_at': '', 'volume_id': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2', 'serial': '0c2dd27e-67d6-437e-bebb-4a9faa22c7d2'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1257.275085] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1257.275333] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1257.276843] env[62814]: INFO nova.compute.claims [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1257.459186] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294828, 'name': Rename_Task, 'duration_secs': 0.132507} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.461438] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1257.461687] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3f7be42-0d95-4862-9933-ae9a04217d72 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.467461] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1257.467461] env[62814]: value = "task-4294829" [ 1257.467461] env[62814]: _type = "Task" [ 1257.467461] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.474586] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294829, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.515340] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52634ffe-027e-2fd7-cbe2-d9b3483fe7bd, 'name': SearchDatastore_Task, 'duration_secs': 0.007976} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.516109] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-809ce566-68ee-403c-ac6a-37a28f10e7bf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.520820] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1257.520820] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52943485-1bc2-40de-e036-cd1c1c24de67" [ 1257.520820] env[62814]: _type = "Task" [ 1257.520820] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1257.529739] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52943485-1bc2-40de-e036-cd1c1c24de67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1257.592073] env[62814]: DEBUG nova.network.neutron [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance_info_cache with network_info: [{"id": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "address": "fa:16:3e:94:da:3d", "network": {"id": "9d2d41c6-d393-4446-a759-e14d1d871d3e", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1627222555-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6e52480dd2c467790622901940cf385", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "51bac3c3-00ab-4a07-9e28-b3c951dee565", "external-id": "nsx-vlan-transportzone-645", "segmentation_id": 645, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d17bc61-a2", "ovs_interfaceid": "3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1257.827883] env[62814]: DEBUG nova.objects.instance [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lazy-loading 'flavor' on Instance uuid 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1257.981076] env[62814]: DEBUG oslo_vmware.api [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294829, 'name': PowerOnVM_Task, 'duration_secs': 0.434388} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1257.981429] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1257.981571] env[62814]: INFO nova.compute.manager [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Took 6.81 seconds to spawn the instance on the hypervisor. [ 1257.981753] env[62814]: DEBUG nova.compute.manager [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1257.982601] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a708215-9815-457a-8d0e-0ebbddc04f97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.030825] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52943485-1bc2-40de-e036-cd1c1c24de67, 'name': SearchDatastore_Task, 'duration_secs': 0.010905} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.031134] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1258.031394] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] add41cff-326f-46f4-b768-1f08ace2b31f/add41cff-326f-46f4-b768-1f08ace2b31f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1258.031660] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc8e21a6-da09-45b5-a941-80dcbcfd3070 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.039327] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1258.039327] env[62814]: value = "task-4294830" [ 1258.039327] env[62814]: _type = "Task" [ 1258.039327] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.047547] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.095101] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Releasing lock "refresh_cache-ae9180b8-6caa-44ef-93a8-eb3a7681d224" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1258.095343] env[62814]: DEBUG nova.objects.instance [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lazy-loading 'migration_context' on Instance uuid ae9180b8-6caa-44ef-93a8-eb3a7681d224 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1258.408842] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de44b7e-b92d-4c17-abac-4a2ec624e7e4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.417488] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85fa4548-ed0f-4bab-98ef-d7d6022902c2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.450791] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da08d4d-fe51-49d8-8718-af764bb47da1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.458548] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d802436-d683-4778-bbaf-280f3076e22e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.472553] env[62814]: DEBUG nova.compute.provider_tree [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1258.498326] env[62814]: INFO nova.compute.manager [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Took 11.59 seconds to build instance. [ 1258.548217] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294830, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464038} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.548478] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] add41cff-326f-46f4-b768-1f08ace2b31f/add41cff-326f-46f4-b768-1f08ace2b31f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1258.548695] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1258.548940] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2a000c0a-3fdf-440e-a22a-403bb61c9a90 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.554658] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1258.554658] env[62814]: value = "task-4294831" [ 1258.554658] env[62814]: _type = "Task" [ 1258.554658] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.562814] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294831, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1258.598668] env[62814]: DEBUG nova.objects.base [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1258.599583] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74d6c44a-d998-4480-8cac-c3529fcc48cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.618231] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-400ba215-4add-496a-94f9-9d3521528721 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.623051] env[62814]: DEBUG oslo_vmware.api [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1258.623051] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e8906b-5548-6ca5-5997-9349da3785bc" [ 1258.623051] env[62814]: _type = "Task" [ 1258.623051] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.630622] env[62814]: DEBUG oslo_vmware.api [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e8906b-5548-6ca5-5997-9349da3785bc, 'name': SearchDatastore_Task, 'duration_secs': 0.00558} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.630879] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1258.836919] env[62814]: DEBUG oslo_concurrency.lockutils [None req-809ddac7-cf0d-4657-928b-86321272d96f tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.241s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1258.976521] env[62814]: DEBUG nova.scheduler.client.report [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1259.000180] env[62814]: DEBUG oslo_concurrency.lockutils [None req-eb5d17b2-f7c9-4199-b5e5-9c53a4e6d340 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.095s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1259.064525] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294831, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055872} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.064877] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1259.065699] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7717d2-006f-4b35-8654-674e51ac5637 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.087649] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Reconfiguring VM instance instance-00000078 to attach disk [datastore2] add41cff-326f-46f4-b768-1f08ace2b31f/add41cff-326f-46f4-b768-1f08ace2b31f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1259.087939] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad4118c9-9264-481e-808d-c30aaf939744 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.108799] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1259.108799] env[62814]: value = "task-4294832" [ 1259.108799] env[62814]: _type = "Task" [ 1259.108799] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.116305] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294832, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.481321] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1259.481847] env[62814]: DEBUG nova.compute.manager [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1259.485156] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.854s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1259.618882] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294832, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.766424] env[62814]: DEBUG nova.compute.manager [req-77d0456f-cdec-4fa9-9a10-6b4588c8c448 req-f02176e4-f92a-4a36-a60a-8847134f0a2b service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received event network-changed-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1259.766642] env[62814]: DEBUG nova.compute.manager [req-77d0456f-cdec-4fa9-9a10-6b4588c8c448 req-f02176e4-f92a-4a36-a60a-8847134f0a2b service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Refreshing instance network info cache due to event network-changed-d584963d-9840-4607-b75d-b6a46b94cea3. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1259.766860] env[62814]: DEBUG oslo_concurrency.lockutils [req-77d0456f-cdec-4fa9-9a10-6b4588c8c448 req-f02176e4-f92a-4a36-a60a-8847134f0a2b service nova] Acquiring lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1259.767097] env[62814]: DEBUG oslo_concurrency.lockutils [req-77d0456f-cdec-4fa9-9a10-6b4588c8c448 req-f02176e4-f92a-4a36-a60a-8847134f0a2b service nova] Acquired lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1259.767201] env[62814]: DEBUG nova.network.neutron [req-77d0456f-cdec-4fa9-9a10-6b4588c8c448 req-f02176e4-f92a-4a36-a60a-8847134f0a2b service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Refreshing network info cache for port d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1259.856067] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1259.856067] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1259.856067] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1259.856067] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1259.856388] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1259.858435] env[62814]: INFO nova.compute.manager [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Terminating instance [ 1259.988678] env[62814]: DEBUG nova.compute.utils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1259.990142] env[62814]: DEBUG nova.compute.manager [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1259.990331] env[62814]: DEBUG nova.network.neutron [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1260.034838] env[62814]: DEBUG nova.policy [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ca28a465227468a9f21fa208f16ca7c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ed3e3bdfb4d4646a8c9eef1582cde85', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1260.103449] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8794dd9c-6f6f-49e8-bd04-033b9e72fc47 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.114037] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6aa723d-0831-447b-a949-4a7e7972180a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.121566] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294832, 'name': ReconfigVM_Task, 'duration_secs': 0.531395} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.147506] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Reconfigured VM instance instance-00000078 to attach disk [datastore2] add41cff-326f-46f4-b768-1f08ace2b31f/add41cff-326f-46f4-b768-1f08ace2b31f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1260.148825] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f25b287d-abcd-43bc-9df8-f2b595722527 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.151016] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a6f589f-1d58-4d02-9dda-caaddbe84f69 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.160778] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c728b4-244b-4f24-945c-287aebc3b520 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.164673] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1260.164673] env[62814]: value = "task-4294833" [ 1260.164673] env[62814]: _type = "Task" [ 1260.164673] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.177896] env[62814]: DEBUG nova.compute.provider_tree [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1260.184115] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294833, 'name': Rename_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.342288] env[62814]: DEBUG nova.network.neutron [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Successfully created port: b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1260.362193] env[62814]: DEBUG nova.compute.manager [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1260.362447] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1260.363385] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a031d3-68be-4390-9c3c-1e3e34cf372a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.371534] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1260.374463] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-39d38ccb-41fd-40b7-818f-3d5c21afbe2d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.381508] env[62814]: DEBUG oslo_vmware.api [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1260.381508] env[62814]: value = "task-4294834" [ 1260.381508] env[62814]: _type = "Task" [ 1260.381508] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.390368] env[62814]: DEBUG oslo_vmware.api [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294834, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.496714] env[62814]: DEBUG nova.compute.manager [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1260.501117] env[62814]: DEBUG nova.network.neutron [req-77d0456f-cdec-4fa9-9a10-6b4588c8c448 req-f02176e4-f92a-4a36-a60a-8847134f0a2b service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updated VIF entry in instance network info cache for port d584963d-9840-4607-b75d-b6a46b94cea3. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1260.501483] env[62814]: DEBUG nova.network.neutron [req-77d0456f-cdec-4fa9-9a10-6b4588c8c448 req-f02176e4-f92a-4a36-a60a-8847134f0a2b service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updating instance_info_cache with network_info: [{"id": "d584963d-9840-4607-b75d-b6a46b94cea3", "address": "fa:16:3e:ca:e5:53", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd584963d-98", "ovs_interfaceid": "d584963d-9840-4607-b75d-b6a46b94cea3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1260.675409] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294833, 'name': Rename_Task, 'duration_secs': 0.256334} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.675734] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1260.676007] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67b68811-0f20-4712-8534-000fbe174be5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.683606] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1260.683606] env[62814]: value = "task-4294835" [ 1260.683606] env[62814]: _type = "Task" [ 1260.683606] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.687404] env[62814]: DEBUG nova.scheduler.client.report [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1260.695910] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294835, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.891211] env[62814]: DEBUG oslo_vmware.api [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294834, 'name': PowerOffVM_Task, 'duration_secs': 0.273425} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1260.891558] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1260.891832] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1260.892173] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-38f87bdf-de60-494b-b3e9-01fdacae39d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.955612] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1260.955949] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1260.956271] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleting the datastore file [datastore2] 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1260.956622] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5e759539-b80d-4121-a2b6-30fa699587d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1260.963910] env[62814]: DEBUG oslo_vmware.api [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for the task: (returnval){ [ 1260.963910] env[62814]: value = "task-4294837" [ 1260.963910] env[62814]: _type = "Task" [ 1260.963910] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1260.974223] env[62814]: DEBUG oslo_vmware.api [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294837, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.004763] env[62814]: DEBUG oslo_concurrency.lockutils [req-77d0456f-cdec-4fa9-9a10-6b4588c8c448 req-f02176e4-f92a-4a36-a60a-8847134f0a2b service nova] Releasing lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1261.197162] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294835, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.474018] env[62814]: DEBUG oslo_vmware.api [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Task: {'id': task-4294837, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133276} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.474306] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1261.474489] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1261.474690] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1261.474857] env[62814]: INFO nova.compute.manager [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1261.475145] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1261.475354] env[62814]: DEBUG nova.compute.manager [-] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1261.475451] env[62814]: DEBUG nova.network.neutron [-] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1261.513816] env[62814]: DEBUG nova.compute.manager [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1261.550290] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1261.550618] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1261.550841] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1261.551101] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1261.551348] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1261.551617] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1261.551855] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1261.552032] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1261.552217] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1261.552387] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1261.552563] env[62814]: DEBUG nova.virt.hardware [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1261.553478] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bba71b-9b81-4e51-b288-c1699933657b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.562683] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80cc7eb7-3c2d-457a-86ae-7f8fead2fd2d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.700534] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.214s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1261.705455] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294835, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.851222] env[62814]: DEBUG nova.compute.manager [req-6dfab10c-a542-49cc-b946-58f16654cb1f req-70f026e7-a2a8-490e-951a-ba9cc0ae9b62 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received event network-vif-plugged-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1261.851222] env[62814]: DEBUG oslo_concurrency.lockutils [req-6dfab10c-a542-49cc-b946-58f16654cb1f req-70f026e7-a2a8-490e-951a-ba9cc0ae9b62 service nova] Acquiring lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1261.851222] env[62814]: DEBUG oslo_concurrency.lockutils [req-6dfab10c-a542-49cc-b946-58f16654cb1f req-70f026e7-a2a8-490e-951a-ba9cc0ae9b62 service nova] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1261.851222] env[62814]: DEBUG oslo_concurrency.lockutils [req-6dfab10c-a542-49cc-b946-58f16654cb1f req-70f026e7-a2a8-490e-951a-ba9cc0ae9b62 service nova] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1261.851222] env[62814]: DEBUG nova.compute.manager [req-6dfab10c-a542-49cc-b946-58f16654cb1f req-70f026e7-a2a8-490e-951a-ba9cc0ae9b62 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] No waiting events found dispatching network-vif-plugged-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1261.851222] env[62814]: WARNING nova.compute.manager [req-6dfab10c-a542-49cc-b946-58f16654cb1f req-70f026e7-a2a8-490e-951a-ba9cc0ae9b62 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received unexpected event network-vif-plugged-b5888ff4-67ff-4165-a2bc-390d621e4efa for instance with vm_state building and task_state spawning. [ 1261.855092] env[62814]: DEBUG nova.network.neutron [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Successfully updated port: b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1262.018311] env[62814]: DEBUG nova.compute.manager [req-ccba248d-36f3-4133-9f4d-4cb973da578e req-58429f3d-8e3d-41f6-8c51-70f04eae227e service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Received event network-vif-deleted-e07ce73b-c8ba-4a79-a9a4-a5d1672041d9 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1262.018497] env[62814]: INFO nova.compute.manager [req-ccba248d-36f3-4133-9f4d-4cb973da578e req-58429f3d-8e3d-41f6-8c51-70f04eae227e service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Neutron deleted interface e07ce73b-c8ba-4a79-a9a4-a5d1672041d9; detaching it from the instance and deleting it from the info cache [ 1262.018667] env[62814]: DEBUG nova.network.neutron [req-ccba248d-36f3-4133-9f4d-4cb973da578e req-58429f3d-8e3d-41f6-8c51-70f04eae227e service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1262.195494] env[62814]: DEBUG oslo_vmware.api [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294835, 'name': PowerOnVM_Task, 'duration_secs': 1.163271} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.195754] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1262.195952] env[62814]: INFO nova.compute.manager [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Took 8.82 seconds to spawn the instance on the hypervisor. [ 1262.196150] env[62814]: DEBUG nova.compute.manager [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1262.196979] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54abd9b8-903c-40a5-ba89-9e7ff4ea6358 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.262400] env[62814]: INFO nova.scheduler.client.report [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted allocation for migration e0d9dda5-4f55-449b-b60e-95c12717178c [ 1262.358445] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1262.358623] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1262.358759] env[62814]: DEBUG nova.network.neutron [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1262.500103] env[62814]: DEBUG nova.network.neutron [-] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1262.521575] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e6a4105b-0c7a-47f5-9c4a-34fcbb26c0b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.532052] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78e26ea6-6425-4715-bdbc-5ea0983f6e4a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.559291] env[62814]: DEBUG nova.compute.manager [req-ccba248d-36f3-4133-9f4d-4cb973da578e req-58429f3d-8e3d-41f6-8c51-70f04eae227e service nova] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Detach interface failed, port_id=e07ce73b-c8ba-4a79-a9a4-a5d1672041d9, reason: Instance 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1262.718195] env[62814]: INFO nova.compute.manager [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Took 13.78 seconds to build instance. [ 1262.768463] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.928s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1262.899741] env[62814]: DEBUG nova.network.neutron [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1263.003529] env[62814]: INFO nova.compute.manager [-] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Took 1.53 seconds to deallocate network for instance. [ 1263.087681] env[62814]: DEBUG nova.network.neutron [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updating instance_info_cache with network_info: [{"id": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "address": "fa:16:3e:02:56:99", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5888ff4-67", "ovs_interfaceid": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1263.219557] env[62814]: DEBUG oslo_concurrency.lockutils [None req-5c5f9ae6-f678-4bfc-b4eb-f237686324c0 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "add41cff-326f-46f4-b768-1f08ace2b31f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.291s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1263.511569] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1263.511569] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1263.511569] env[62814]: DEBUG nova.objects.instance [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lazy-loading 'resources' on Instance uuid 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1263.590783] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1263.591069] env[62814]: DEBUG nova.compute.manager [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Instance network_info: |[{"id": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "address": "fa:16:3e:02:56:99", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5888ff4-67", "ovs_interfaceid": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1263.591783] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:56:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5888ff4-67ff-4165-a2bc-390d621e4efa', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1263.601012] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1263.601409] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1263.601722] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fbe257b9-e327-4536-9e53-b8d8e5362db1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.622400] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1263.622400] env[62814]: value = "task-4294838" [ 1263.622400] env[62814]: _type = "Task" [ 1263.622400] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.630501] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294838, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.878957] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1263.878957] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1263.878957] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1263.879158] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1263.879290] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1263.882894] env[62814]: DEBUG nova.compute.manager [req-3a69df8e-2cb6-4e15-bd59-7960f184dd24 req-6eb6f68d-6a01-4577-abd7-a3e019cefe76 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received event network-changed-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1263.883241] env[62814]: DEBUG nova.compute.manager [req-3a69df8e-2cb6-4e15-bd59-7960f184dd24 req-6eb6f68d-6a01-4577-abd7-a3e019cefe76 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Refreshing instance network info cache due to event network-changed-b5888ff4-67ff-4165-a2bc-390d621e4efa. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1263.883455] env[62814]: DEBUG oslo_concurrency.lockutils [req-3a69df8e-2cb6-4e15-bd59-7960f184dd24 req-6eb6f68d-6a01-4577-abd7-a3e019cefe76 service nova] Acquiring lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1263.883597] env[62814]: DEBUG oslo_concurrency.lockutils [req-3a69df8e-2cb6-4e15-bd59-7960f184dd24 req-6eb6f68d-6a01-4577-abd7-a3e019cefe76 service nova] Acquired lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1263.883759] env[62814]: DEBUG nova.network.neutron [req-3a69df8e-2cb6-4e15-bd59-7960f184dd24 req-6eb6f68d-6a01-4577-abd7-a3e019cefe76 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Refreshing network info cache for port b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1263.885109] env[62814]: INFO nova.compute.manager [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Terminating instance [ 1264.051951] env[62814]: DEBUG nova.compute.manager [req-7908ff51-dbbf-4860-882b-2dadb9f48036 req-66f2383e-7712-49ea-9052-b8794ff39e56 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Received event network-changed-14c4e34d-9d53-41a8-ade9-86a6fa15cab5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1264.052176] env[62814]: DEBUG nova.compute.manager [req-7908ff51-dbbf-4860-882b-2dadb9f48036 req-66f2383e-7712-49ea-9052-b8794ff39e56 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Refreshing instance network info cache due to event network-changed-14c4e34d-9d53-41a8-ade9-86a6fa15cab5. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1264.052408] env[62814]: DEBUG oslo_concurrency.lockutils [req-7908ff51-dbbf-4860-882b-2dadb9f48036 req-66f2383e-7712-49ea-9052-b8794ff39e56 service nova] Acquiring lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.052559] env[62814]: DEBUG oslo_concurrency.lockutils [req-7908ff51-dbbf-4860-882b-2dadb9f48036 req-66f2383e-7712-49ea-9052-b8794ff39e56 service nova] Acquired lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1264.052757] env[62814]: DEBUG nova.network.neutron [req-7908ff51-dbbf-4860-882b-2dadb9f48036 req-66f2383e-7712-49ea-9052-b8794ff39e56 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Refreshing network info cache for port 14c4e34d-9d53-41a8-ade9-86a6fa15cab5 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1264.112161] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f96b15-16ad-4046-b650-0aa8d88319ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.119639] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c225d3a9-e022-4722-8cdc-c5daaf71ded3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.130804] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294838, 'name': CreateVM_Task, 'duration_secs': 0.281312} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.153626] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1264.154733] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.154911] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1264.155249] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1264.155978] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aec4254-e6e8-4bf9-a641-40674a3c468c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.158292] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04f443a3-42b6-4f1d-b960-906dab720a91 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.165221] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b5656d-f73e-47eb-bb30-05eb35a07073 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.168819] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1264.168819] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cf36b9-9a31-dd41-4042-ee7fae70eb4d" [ 1264.168819] env[62814]: _type = "Task" [ 1264.168819] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.180963] env[62814]: DEBUG nova.compute.provider_tree [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1264.187074] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cf36b9-9a31-dd41-4042-ee7fae70eb4d, 'name': SearchDatastore_Task, 'duration_secs': 0.010394} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.187738] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1264.188686] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1264.188686] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.188686] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1264.188686] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1264.188686] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d11d7a43-3f4b-42b6-b575-4c685c916d7a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.196405] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1264.196578] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1264.197262] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85c7cb2c-baf9-4cdc-856b-a3d987759695 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.202302] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1264.202302] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc3126-d40e-6fc2-0cfa-3b8d3a0e58db" [ 1264.202302] env[62814]: _type = "Task" [ 1264.202302] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.209236] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc3126-d40e-6fc2-0cfa-3b8d3a0e58db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.391844] env[62814]: DEBUG nova.compute.manager [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1264.393010] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1264.393218] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb2ba660-2cfc-4d4d-9625-554c20e9eb00 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.402428] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1264.402730] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42db01c2-f077-4857-8880-3688309e7ab0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.409145] env[62814]: DEBUG oslo_vmware.api [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1264.409145] env[62814]: value = "task-4294839" [ 1264.409145] env[62814]: _type = "Task" [ 1264.409145] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.419336] env[62814]: DEBUG oslo_vmware.api [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294839, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.646525] env[62814]: DEBUG nova.network.neutron [req-3a69df8e-2cb6-4e15-bd59-7960f184dd24 req-6eb6f68d-6a01-4577-abd7-a3e019cefe76 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updated VIF entry in instance network info cache for port b5888ff4-67ff-4165-a2bc-390d621e4efa. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1264.646525] env[62814]: DEBUG nova.network.neutron [req-3a69df8e-2cb6-4e15-bd59-7960f184dd24 req-6eb6f68d-6a01-4577-abd7-a3e019cefe76 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updating instance_info_cache with network_info: [{"id": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "address": "fa:16:3e:02:56:99", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5888ff4-67", "ovs_interfaceid": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.683522] env[62814]: DEBUG nova.scheduler.client.report [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1264.715884] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cc3126-d40e-6fc2-0cfa-3b8d3a0e58db, 'name': SearchDatastore_Task, 'duration_secs': 0.008124} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.716681] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d15c839-6413-4636-abae-eae66ddd4ba2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.721769] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1264.721769] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5277e089-f459-e5e0-ef5c-bd21369a3789" [ 1264.721769] env[62814]: _type = "Task" [ 1264.721769] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.729210] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5277e089-f459-e5e0-ef5c-bd21369a3789, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.771929] env[62814]: DEBUG nova.network.neutron [req-7908ff51-dbbf-4860-882b-2dadb9f48036 req-66f2383e-7712-49ea-9052-b8794ff39e56 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Updated VIF entry in instance network info cache for port 14c4e34d-9d53-41a8-ade9-86a6fa15cab5. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1264.772498] env[62814]: DEBUG nova.network.neutron [req-7908ff51-dbbf-4860-882b-2dadb9f48036 req-66f2383e-7712-49ea-9052-b8794ff39e56 service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Updating instance_info_cache with network_info: [{"id": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "address": "fa:16:3e:c3:d9:f1", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14c4e34d-9d", "ovs_interfaceid": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.919346] env[62814]: DEBUG oslo_vmware.api [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294839, 'name': PowerOffVM_Task, 'duration_secs': 0.173927} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1264.919616] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1264.919783] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1264.920051] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e762095b-abb0-456e-b336-ca8ed2c97409 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.977599] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1264.977947] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1264.978159] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleting the datastore file [datastore2] ae9180b8-6caa-44ef-93a8-eb3a7681d224 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1264.978423] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a5f4fdf9-b3a6-4997-9980-a978d1375c4e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1264.984961] env[62814]: DEBUG oslo_vmware.api [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for the task: (returnval){ [ 1264.984961] env[62814]: value = "task-4294841" [ 1264.984961] env[62814]: _type = "Task" [ 1264.984961] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1264.992817] env[62814]: DEBUG oslo_vmware.api [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294841, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.148879] env[62814]: DEBUG oslo_concurrency.lockutils [req-3a69df8e-2cb6-4e15-bd59-7960f184dd24 req-6eb6f68d-6a01-4577-abd7-a3e019cefe76 service nova] Releasing lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1265.189191] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.678s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1265.207555] env[62814]: INFO nova.scheduler.client.report [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Deleted allocations for instance 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac [ 1265.233593] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5277e089-f459-e5e0-ef5c-bd21369a3789, 'name': SearchDatastore_Task, 'duration_secs': 0.02846} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.233797] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1265.234223] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] ba933282-9711-4e99-ad2d-6ad4c6c516cc/ba933282-9711-4e99-ad2d-6ad4c6c516cc.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1265.234427] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af7643b4-1a8e-49b7-8445-8f86998694d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.241309] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1265.241309] env[62814]: value = "task-4294842" [ 1265.241309] env[62814]: _type = "Task" [ 1265.241309] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.248566] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294842, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.275483] env[62814]: DEBUG oslo_concurrency.lockutils [req-7908ff51-dbbf-4860-882b-2dadb9f48036 req-66f2383e-7712-49ea-9052-b8794ff39e56 service nova] Releasing lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1265.494957] env[62814]: DEBUG oslo_vmware.api [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Task: {'id': task-4294841, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.35925} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.495262] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1265.495510] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1265.495754] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1265.495950] env[62814]: INFO nova.compute.manager [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1265.496218] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1265.496410] env[62814]: DEBUG nova.compute.manager [-] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1265.496506] env[62814]: DEBUG nova.network.neutron [-] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1265.716185] env[62814]: DEBUG oslo_concurrency.lockutils [None req-62bd7b25-225f-4a17-9b37-7171baca8104 tempest-AttachVolumeNegativeTest-1633709727 tempest-AttachVolumeNegativeTest-1633709727-project-member] Lock "0fdb57c1-7a2a-455d-acb8-9f342ef1dbac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.860s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1265.755961] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294842, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.914257] env[62814]: DEBUG nova.compute.manager [req-718ee61b-0ca8-4791-904b-23d73409b945 req-4861ded2-d2c0-4845-bffe-2ee8f5ac6e6c service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Received event network-vif-deleted-3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1265.914567] env[62814]: INFO nova.compute.manager [req-718ee61b-0ca8-4791-904b-23d73409b945 req-4861ded2-d2c0-4845-bffe-2ee8f5ac6e6c service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Neutron deleted interface 3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8; detaching it from the instance and deleting it from the info cache [ 1265.914642] env[62814]: DEBUG nova.network.neutron [req-718ee61b-0ca8-4791-904b-23d73409b945 req-4861ded2-d2c0-4845-bffe-2ee8f5ac6e6c service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1266.251610] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294842, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.751799} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.251929] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] ba933282-9711-4e99-ad2d-6ad4c6c516cc/ba933282-9711-4e99-ad2d-6ad4c6c516cc.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1266.252165] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1266.252411] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83a4447e-6e06-4f53-99d0-8aae82e1e02b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.259212] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1266.259212] env[62814]: value = "task-4294843" [ 1266.259212] env[62814]: _type = "Task" [ 1266.259212] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.266156] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.303630] env[62814]: DEBUG nova.network.neutron [-] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1266.417436] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-31f59f86-0c14-42d1-83f3-b3ee86501fef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.426510] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f3cf05-567f-46ca-8c49-1547db99237f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.455822] env[62814]: DEBUG nova.compute.manager [req-718ee61b-0ca8-4791-904b-23d73409b945 req-4861ded2-d2c0-4845-bffe-2ee8f5ac6e6c service nova] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Detach interface failed, port_id=3d17bc61-a2a9-4a92-bb81-6e4c3f2cd4a8, reason: Instance ae9180b8-6caa-44ef-93a8-eb3a7681d224 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1266.648481] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "interface-b5a2fb6c-7078-4f3c-b511-014beb96391d-d8936f3d-f722-43da-bf8e-17c76e3d95eb" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1266.648767] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-b5a2fb6c-7078-4f3c-b511-014beb96391d-d8936f3d-f722-43da-bf8e-17c76e3d95eb" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1266.649168] env[62814]: DEBUG nova.objects.instance [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'flavor' on Instance uuid b5a2fb6c-7078-4f3c-b511-014beb96391d {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1266.769248] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067738} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.769646] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1266.770261] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6253913-99b7-4a41-8d90-f5de6a77b0a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.793638] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Reconfiguring VM instance instance-00000079 to attach disk [datastore2] ba933282-9711-4e99-ad2d-6ad4c6c516cc/ba933282-9711-4e99-ad2d-6ad4c6c516cc.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1266.793904] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9717150d-6f9b-470f-aaca-e2468411d729 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.808195] env[62814]: INFO nova.compute.manager [-] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Took 1.31 seconds to deallocate network for instance. [ 1266.814913] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1266.814913] env[62814]: value = "task-4294845" [ 1266.814913] env[62814]: _type = "Task" [ 1266.814913] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.822891] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294845, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.254529] env[62814]: DEBUG nova.objects.instance [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'pci_requests' on Instance uuid b5a2fb6c-7078-4f3c-b511-014beb96391d {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1267.317903] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1267.318381] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1267.318669] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1267.334623] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294845, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.342675] env[62814]: INFO nova.scheduler.client.report [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Deleted allocations for instance ae9180b8-6caa-44ef-93a8-eb3a7681d224 [ 1267.757553] env[62814]: DEBUG nova.objects.base [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1267.757790] env[62814]: DEBUG nova.network.neutron [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1267.831218] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294845, 'name': ReconfigVM_Task, 'duration_secs': 0.66742} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.832201] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Reconfigured VM instance instance-00000079 to attach disk [datastore2] ba933282-9711-4e99-ad2d-6ad4c6c516cc/ba933282-9711-4e99-ad2d-6ad4c6c516cc.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1267.832201] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-14b4eb54-135d-48dd-88b4-ca499f01e49d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.838527] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1267.838527] env[62814]: value = "task-4294846" [ 1267.838527] env[62814]: _type = "Task" [ 1267.838527] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.839912] env[62814]: DEBUG nova.policy [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5779eaa60ec44b2b80660e521b598eb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10018060297c409d82935ad626cc54c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1267.851404] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294846, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.851846] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9548a1fc-f26a-451e-8c95-23f46bd60cd2 tempest-DeleteServersTestJSON-1537205367 tempest-DeleteServersTestJSON-1537205367-project-member] Lock "ae9180b8-6caa-44ef-93a8-eb3a7681d224" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.973s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1268.353121] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294846, 'name': Rename_Task, 'duration_secs': 0.146841} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.353658] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1268.353942] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7b03f526-8395-4fb8-8511-a2fc4aa22ab1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.360764] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1268.360764] env[62814]: value = "task-4294847" [ 1268.360764] env[62814]: _type = "Task" [ 1268.360764] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1268.368281] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294847, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.870013] env[62814]: DEBUG oslo_vmware.api [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294847, 'name': PowerOnVM_Task, 'duration_secs': 0.431144} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.870344] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1268.870495] env[62814]: INFO nova.compute.manager [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Took 7.36 seconds to spawn the instance on the hypervisor. [ 1268.870671] env[62814]: DEBUG nova.compute.manager [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1268.871434] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec9e068-0a2c-42ad-b266-dd724ab33f86 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.264565] env[62814]: DEBUG nova.compute.manager [req-fc41493b-1294-440e-a4b9-9b2ebbdd3ca6 req-9810e5aa-2d0a-46d5-844e-8e93a90407fb service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received event network-vif-plugged-d8936f3d-f722-43da-bf8e-17c76e3d95eb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1269.264821] env[62814]: DEBUG oslo_concurrency.lockutils [req-fc41493b-1294-440e-a4b9-9b2ebbdd3ca6 req-9810e5aa-2d0a-46d5-844e-8e93a90407fb service nova] Acquiring lock "b5a2fb6c-7078-4f3c-b511-014beb96391d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1269.265050] env[62814]: DEBUG oslo_concurrency.lockutils [req-fc41493b-1294-440e-a4b9-9b2ebbdd3ca6 req-9810e5aa-2d0a-46d5-844e-8e93a90407fb service nova] Lock "b5a2fb6c-7078-4f3c-b511-014beb96391d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1269.265332] env[62814]: DEBUG oslo_concurrency.lockutils [req-fc41493b-1294-440e-a4b9-9b2ebbdd3ca6 req-9810e5aa-2d0a-46d5-844e-8e93a90407fb service nova] Lock "b5a2fb6c-7078-4f3c-b511-014beb96391d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1269.265395] env[62814]: DEBUG nova.compute.manager [req-fc41493b-1294-440e-a4b9-9b2ebbdd3ca6 req-9810e5aa-2d0a-46d5-844e-8e93a90407fb service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] No waiting events found dispatching network-vif-plugged-d8936f3d-f722-43da-bf8e-17c76e3d95eb {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1269.265597] env[62814]: WARNING nova.compute.manager [req-fc41493b-1294-440e-a4b9-9b2ebbdd3ca6 req-9810e5aa-2d0a-46d5-844e-8e93a90407fb service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received unexpected event network-vif-plugged-d8936f3d-f722-43da-bf8e-17c76e3d95eb for instance with vm_state active and task_state None. [ 1269.377671] env[62814]: DEBUG nova.network.neutron [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Successfully updated port: d8936f3d-f722-43da-bf8e-17c76e3d95eb {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1269.389303] env[62814]: INFO nova.compute.manager [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Took 12.13 seconds to build instance. [ 1269.782447] env[62814]: DEBUG nova.compute.manager [req-f8f472fc-0954-47bf-a9e0-615164b6d4e1 req-5f446f60-c2bb-444d-bbed-5fd7304c3fb4 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received event network-changed-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1269.782653] env[62814]: DEBUG nova.compute.manager [req-f8f472fc-0954-47bf-a9e0-615164b6d4e1 req-5f446f60-c2bb-444d-bbed-5fd7304c3fb4 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Refreshing instance network info cache due to event network-changed-b5888ff4-67ff-4165-a2bc-390d621e4efa. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1269.782870] env[62814]: DEBUG oslo_concurrency.lockutils [req-f8f472fc-0954-47bf-a9e0-615164b6d4e1 req-5f446f60-c2bb-444d-bbed-5fd7304c3fb4 service nova] Acquiring lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1269.783031] env[62814]: DEBUG oslo_concurrency.lockutils [req-f8f472fc-0954-47bf-a9e0-615164b6d4e1 req-5f446f60-c2bb-444d-bbed-5fd7304c3fb4 service nova] Acquired lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1269.783204] env[62814]: DEBUG nova.network.neutron [req-f8f472fc-0954-47bf-a9e0-615164b6d4e1 req-5f446f60-c2bb-444d-bbed-5fd7304c3fb4 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Refreshing network info cache for port b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1269.886063] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1269.886063] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1269.886063] env[62814]: DEBUG nova.network.neutron [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1269.891601] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4db74f8e-31c2-4930-b8ab-f20ca3250ca4 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.657s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1270.433280] env[62814]: WARNING nova.network.neutron [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] 1de03cb0-69fa-4200-8fa1-26375f44369a already exists in list: networks containing: ['1de03cb0-69fa-4200-8fa1-26375f44369a']. ignoring it [ 1270.509348] env[62814]: DEBUG nova.network.neutron [req-f8f472fc-0954-47bf-a9e0-615164b6d4e1 req-5f446f60-c2bb-444d-bbed-5fd7304c3fb4 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updated VIF entry in instance network info cache for port b5888ff4-67ff-4165-a2bc-390d621e4efa. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1270.509938] env[62814]: DEBUG nova.network.neutron [req-f8f472fc-0954-47bf-a9e0-615164b6d4e1 req-5f446f60-c2bb-444d-bbed-5fd7304c3fb4 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updating instance_info_cache with network_info: [{"id": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "address": "fa:16:3e:02:56:99", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5888ff4-67", "ovs_interfaceid": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1270.699967] env[62814]: DEBUG nova.network.neutron [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [{"id": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "address": "fa:16:3e:ff:ce:8d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9edabfa8-97", "ovs_interfaceid": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d8936f3d-f722-43da-bf8e-17c76e3d95eb", "address": "fa:16:3e:c7:14:77", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8936f3d-f7", "ovs_interfaceid": "d8936f3d-f722-43da-bf8e-17c76e3d95eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1271.015632] env[62814]: DEBUG oslo_concurrency.lockutils [req-f8f472fc-0954-47bf-a9e0-615164b6d4e1 req-5f446f60-c2bb-444d-bbed-5fd7304c3fb4 service nova] Releasing lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1271.203998] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1271.204511] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1271.204825] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1271.205864] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c95ceb8d-3bf9-4460-8e3c-53dc54ec9a3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.225975] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1271.226335] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1271.226553] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1271.226782] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1271.226959] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1271.227150] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1271.227367] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1271.227531] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1271.227702] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1271.227871] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1271.228060] env[62814]: DEBUG nova.virt.hardware [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1271.234400] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Reconfiguring VM to attach interface {{(pid=62814) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1271.235131] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1426cd3a-a245-42c5-8630-77007c63491e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.252668] env[62814]: DEBUG oslo_vmware.api [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1271.252668] env[62814]: value = "task-4294851" [ 1271.252668] env[62814]: _type = "Task" [ 1271.252668] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1271.260939] env[62814]: DEBUG oslo_vmware.api [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294851, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.295228] env[62814]: DEBUG nova.compute.manager [req-6c1b09ea-0492-4751-923c-e013303fc365 req-fda0fbf1-dde0-47d1-b2b5-a759a8671082 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received event network-changed-d8936f3d-f722-43da-bf8e-17c76e3d95eb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1271.295422] env[62814]: DEBUG nova.compute.manager [req-6c1b09ea-0492-4751-923c-e013303fc365 req-fda0fbf1-dde0-47d1-b2b5-a759a8671082 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing instance network info cache due to event network-changed-d8936f3d-f722-43da-bf8e-17c76e3d95eb. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1271.295655] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c1b09ea-0492-4751-923c-e013303fc365 req-fda0fbf1-dde0-47d1-b2b5-a759a8671082 service nova] Acquiring lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1271.295890] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c1b09ea-0492-4751-923c-e013303fc365 req-fda0fbf1-dde0-47d1-b2b5-a759a8671082 service nova] Acquired lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1271.296383] env[62814]: DEBUG nova.network.neutron [req-6c1b09ea-0492-4751-923c-e013303fc365 req-fda0fbf1-dde0-47d1-b2b5-a759a8671082 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing network info cache for port d8936f3d-f722-43da-bf8e-17c76e3d95eb {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1271.764717] env[62814]: DEBUG oslo_vmware.api [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294851, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1272.117197] env[62814]: DEBUG nova.network.neutron [req-6c1b09ea-0492-4751-923c-e013303fc365 req-fda0fbf1-dde0-47d1-b2b5-a759a8671082 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updated VIF entry in instance network info cache for port d8936f3d-f722-43da-bf8e-17c76e3d95eb. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1272.117703] env[62814]: DEBUG nova.network.neutron [req-6c1b09ea-0492-4751-923c-e013303fc365 req-fda0fbf1-dde0-47d1-b2b5-a759a8671082 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [{"id": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "address": "fa:16:3e:ff:ce:8d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9edabfa8-97", "ovs_interfaceid": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d8936f3d-f722-43da-bf8e-17c76e3d95eb", "address": "fa:16:3e:c7:14:77", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8936f3d-f7", "ovs_interfaceid": "d8936f3d-f722-43da-bf8e-17c76e3d95eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1272.263775] env[62814]: DEBUG oslo_vmware.api [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294851, 'name': ReconfigVM_Task, 'duration_secs': 0.672624} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1272.264320] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1272.264575] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Reconfigured VM to attach interface {{(pid=62814) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1272.620527] env[62814]: DEBUG oslo_concurrency.lockutils [req-6c1b09ea-0492-4751-923c-e013303fc365 req-fda0fbf1-dde0-47d1-b2b5-a759a8671082 service nova] Releasing lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1272.769395] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7bab3b1a-a05a-4cf6-ad73-93cbbdbd3e9d tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-b5a2fb6c-7078-4f3c-b511-014beb96391d-d8936f3d-f722-43da-bf8e-17c76e3d95eb" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.120s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1274.384562] env[62814]: DEBUG oslo_concurrency.lockutils [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "interface-b5a2fb6c-7078-4f3c-b511-014beb96391d-d8936f3d-f722-43da-bf8e-17c76e3d95eb" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1274.384843] env[62814]: DEBUG oslo_concurrency.lockutils [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-b5a2fb6c-7078-4f3c-b511-014beb96391d-d8936f3d-f722-43da-bf8e-17c76e3d95eb" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1274.888783] env[62814]: DEBUG oslo_concurrency.lockutils [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.888783] env[62814]: DEBUG oslo_concurrency.lockutils [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1274.889414] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346742a8-464a-4e7a-8ad7-6ad84c4a628e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.907702] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c30955de-ee92-4af9-b601-6039785d4f1a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.935224] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Reconfiguring VM to detach interface {{(pid=62814) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1274.935567] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b7f7295-cebf-43d2-a9f5-f9a72ea8eaf8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.953856] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1274.953856] env[62814]: value = "task-4294852" [ 1274.953856] env[62814]: _type = "Task" [ 1274.953856] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.961595] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.465396] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.964621] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.464757] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.966270] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.465528] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.966327] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.466982] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.971572] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.469137] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.969760] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.469844] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1280.970262] env[62814]: DEBUG oslo_vmware.api [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294852, 'name': ReconfigVM_Task, 'duration_secs': 5.729546} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1280.970514] env[62814]: DEBUG oslo_concurrency.lockutils [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1280.970725] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Reconfigured VM to detach interface {{(pid=62814) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1282.240722] env[62814]: DEBUG oslo_concurrency.lockutils [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1282.241043] env[62814]: DEBUG oslo_concurrency.lockutils [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1282.241127] env[62814]: DEBUG nova.network.neutron [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1282.928413] env[62814]: INFO nova.network.neutron [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Port d8936f3d-f722-43da-bf8e-17c76e3d95eb from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1282.928787] env[62814]: DEBUG nova.network.neutron [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [{"id": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "address": "fa:16:3e:ff:ce:8d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9edabfa8-97", "ovs_interfaceid": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1283.120373] env[62814]: DEBUG nova.compute.manager [req-c9b26da7-b109-4303-8f9b-b01ecdbdb03a req-88774918-769f-4be2-b969-9d104f285455 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received event network-changed-9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1283.120573] env[62814]: DEBUG nova.compute.manager [req-c9b26da7-b109-4303-8f9b-b01ecdbdb03a req-88774918-769f-4be2-b969-9d104f285455 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing instance network info cache due to event network-changed-9edabfa8-9760-4ef4-adfb-afc6e2200b50. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1283.120965] env[62814]: DEBUG oslo_concurrency.lockutils [req-c9b26da7-b109-4303-8f9b-b01ecdbdb03a req-88774918-769f-4be2-b969-9d104f285455 service nova] Acquiring lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1283.433277] env[62814]: DEBUG oslo_concurrency.lockutils [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1283.435486] env[62814]: DEBUG oslo_concurrency.lockutils [req-c9b26da7-b109-4303-8f9b-b01ecdbdb03a req-88774918-769f-4be2-b969-9d104f285455 service nova] Acquired lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1283.435684] env[62814]: DEBUG nova.network.neutron [req-c9b26da7-b109-4303-8f9b-b01ecdbdb03a req-88774918-769f-4be2-b969-9d104f285455 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Refreshing network info cache for port 9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1283.820568] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "interface-7b82801d-7a0f-4baf-ad2e-566174d3eb35-d8936f3d-f722-43da-bf8e-17c76e3d95eb" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1283.820833] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-7b82801d-7a0f-4baf-ad2e-566174d3eb35-d8936f3d-f722-43da-bf8e-17c76e3d95eb" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1283.821146] env[62814]: DEBUG nova.objects.instance [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'flavor' on Instance uuid 7b82801d-7a0f-4baf-ad2e-566174d3eb35 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1283.938127] env[62814]: DEBUG oslo_concurrency.lockutils [None req-175af642-7a66-449f-8e0d-0d6de9e88ef5 tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-b5a2fb6c-7078-4f3c-b511-014beb96391d-d8936f3d-f722-43da-bf8e-17c76e3d95eb" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.553s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1284.153819] env[62814]: DEBUG nova.network.neutron [req-c9b26da7-b109-4303-8f9b-b01ecdbdb03a req-88774918-769f-4be2-b969-9d104f285455 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updated VIF entry in instance network info cache for port 9edabfa8-9760-4ef4-adfb-afc6e2200b50. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1284.154205] env[62814]: DEBUG nova.network.neutron [req-c9b26da7-b109-4303-8f9b-b01ecdbdb03a req-88774918-769f-4be2-b969-9d104f285455 service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [{"id": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "address": "fa:16:3e:ff:ce:8d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9edabfa8-97", "ovs_interfaceid": "9edabfa8-9760-4ef4-adfb-afc6e2200b50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1284.404729] env[62814]: DEBUG nova.objects.instance [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'pci_requests' on Instance uuid 7b82801d-7a0f-4baf-ad2e-566174d3eb35 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1284.656854] env[62814]: DEBUG oslo_concurrency.lockutils [req-c9b26da7-b109-4303-8f9b-b01ecdbdb03a req-88774918-769f-4be2-b969-9d104f285455 service nova] Releasing lock "refresh_cache-b5a2fb6c-7078-4f3c-b511-014beb96391d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1284.907678] env[62814]: DEBUG nova.objects.base [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Object Instance<7b82801d-7a0f-4baf-ad2e-566174d3eb35> lazy-loaded attributes: flavor,pci_requests {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1284.907874] env[62814]: DEBUG nova.network.neutron [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1284.967962] env[62814]: DEBUG nova.policy [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5779eaa60ec44b2b80660e521b598eb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '10018060297c409d82935ad626cc54c9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1285.156680] env[62814]: DEBUG nova.compute.manager [req-0586a836-7fea-4fd3-8045-c78cc9efab65 req-d1254a7f-9ad7-4d31-a992-079da71f35c8 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Received event network-changed-90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1285.156892] env[62814]: DEBUG nova.compute.manager [req-0586a836-7fea-4fd3-8045-c78cc9efab65 req-d1254a7f-9ad7-4d31-a992-079da71f35c8 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Refreshing instance network info cache due to event network-changed-90c053d0-8e56-42b1-89da-b7acd45ac0fa. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1285.157101] env[62814]: DEBUG oslo_concurrency.lockutils [req-0586a836-7fea-4fd3-8045-c78cc9efab65 req-d1254a7f-9ad7-4d31-a992-079da71f35c8 service nova] Acquiring lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1285.157249] env[62814]: DEBUG oslo_concurrency.lockutils [req-0586a836-7fea-4fd3-8045-c78cc9efab65 req-d1254a7f-9ad7-4d31-a992-079da71f35c8 service nova] Acquired lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1285.157439] env[62814]: DEBUG nova.network.neutron [req-0586a836-7fea-4fd3-8045-c78cc9efab65 req-d1254a7f-9ad7-4d31-a992-079da71f35c8 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Refreshing network info cache for port 90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1285.502494] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1285.506162] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1285.506377] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1285.506525] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 1285.855226] env[62814]: DEBUG nova.network.neutron [req-0586a836-7fea-4fd3-8045-c78cc9efab65 req-d1254a7f-9ad7-4d31-a992-079da71f35c8 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updated VIF entry in instance network info cache for port 90c053d0-8e56-42b1-89da-b7acd45ac0fa. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1285.855606] env[62814]: DEBUG nova.network.neutron [req-0586a836-7fea-4fd3-8045-c78cc9efab65 req-d1254a7f-9ad7-4d31-a992-079da71f35c8 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updating instance_info_cache with network_info: [{"id": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "address": "fa:16:3e:e7:24:4d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c053d0-8e", "ovs_interfaceid": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1286.350071] env[62814]: DEBUG nova.compute.manager [req-46299af1-6ebe-4a32-8ff5-98aa68eadea6 req-4e09079c-cef6-4ba7-9a0d-fcf8a117e8e4 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Received event network-vif-plugged-d8936f3d-f722-43da-bf8e-17c76e3d95eb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1286.350317] env[62814]: DEBUG oslo_concurrency.lockutils [req-46299af1-6ebe-4a32-8ff5-98aa68eadea6 req-4e09079c-cef6-4ba7-9a0d-fcf8a117e8e4 service nova] Acquiring lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1286.350455] env[62814]: DEBUG oslo_concurrency.lockutils [req-46299af1-6ebe-4a32-8ff5-98aa68eadea6 req-4e09079c-cef6-4ba7-9a0d-fcf8a117e8e4 service nova] Lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1286.350656] env[62814]: DEBUG oslo_concurrency.lockutils [req-46299af1-6ebe-4a32-8ff5-98aa68eadea6 req-4e09079c-cef6-4ba7-9a0d-fcf8a117e8e4 service nova] Lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1286.350787] env[62814]: DEBUG nova.compute.manager [req-46299af1-6ebe-4a32-8ff5-98aa68eadea6 req-4e09079c-cef6-4ba7-9a0d-fcf8a117e8e4 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] No waiting events found dispatching network-vif-plugged-d8936f3d-f722-43da-bf8e-17c76e3d95eb {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1286.350949] env[62814]: WARNING nova.compute.manager [req-46299af1-6ebe-4a32-8ff5-98aa68eadea6 req-4e09079c-cef6-4ba7-9a0d-fcf8a117e8e4 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Received unexpected event network-vif-plugged-d8936f3d-f722-43da-bf8e-17c76e3d95eb for instance with vm_state active and task_state None. [ 1286.358084] env[62814]: DEBUG oslo_concurrency.lockutils [req-0586a836-7fea-4fd3-8045-c78cc9efab65 req-d1254a7f-9ad7-4d31-a992-079da71f35c8 service nova] Releasing lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1286.506258] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1286.654629] env[62814]: DEBUG nova.network.neutron [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Successfully updated port: d8936f3d-f722-43da-bf8e-17c76e3d95eb {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1287.009530] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1287.009853] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1287.010050] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1287.010216] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1287.011126] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7a51cf-52b9-4d28-85f8-db91ebc2703f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.019630] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84131ae2-f835-42ca-a03e-9baa193d1de7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.034786] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10286d59-0a14-4e0e-9288-6c6983573177 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.041068] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485333c8-4af2-4131-927e-61b44a9e15d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.070179] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180294MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1287.070496] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1287.070496] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1287.157493] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1287.157949] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1287.157949] env[62814]: DEBUG nova.network.neutron [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1287.693054] env[62814]: WARNING nova.network.neutron [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] 1de03cb0-69fa-4200-8fa1-26375f44369a already exists in list: networks containing: ['1de03cb0-69fa-4200-8fa1-26375f44369a']. ignoring it [ 1287.955761] env[62814]: DEBUG nova.network.neutron [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updating instance_info_cache with network_info: [{"id": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "address": "fa:16:3e:e7:24:4d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c053d0-8e", "ovs_interfaceid": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d8936f3d-f722-43da-bf8e-17c76e3d95eb", "address": "fa:16:3e:c7:14:77", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8936f3d-f7", "ovs_interfaceid": "d8936f3d-f722-43da-bf8e-17c76e3d95eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1288.096661] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance b5a2fb6c-7078-4f3c-b511-014beb96391d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1288.096945] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 7b82801d-7a0f-4baf-ad2e-566174d3eb35 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1288.096945] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance d043b12d-a316-46ab-a30f-abec0a7963a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1288.097059] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance add41cff-326f-46f4-b768-1f08ace2b31f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1288.097168] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance ba933282-9711-4e99-ad2d-6ad4c6c516cc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1288.097349] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1288.097483] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=149GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1288.161127] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6c07c7-f1e5-4fa5-ac43-55243e128fa2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.168760] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0d2d24-945c-4c17-8fbc-bd20961cf0ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.199031] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d589134-bfae-43ee-928e-f99bd9f8ea9d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.205788] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b7e2ac-3b82-4a2c-9610-9ac10c81c6c8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.218586] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1288.376268] env[62814]: DEBUG nova.compute.manager [req-8183e97a-431d-4e1d-8bba-0758530e4f7c req-9ddbb5ca-8564-4d65-9c71-1d6ec8110048 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Received event network-changed-d8936f3d-f722-43da-bf8e-17c76e3d95eb {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1288.376460] env[62814]: DEBUG nova.compute.manager [req-8183e97a-431d-4e1d-8bba-0758530e4f7c req-9ddbb5ca-8564-4d65-9c71-1d6ec8110048 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Refreshing instance network info cache due to event network-changed-d8936f3d-f722-43da-bf8e-17c76e3d95eb. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1288.376647] env[62814]: DEBUG oslo_concurrency.lockutils [req-8183e97a-431d-4e1d-8bba-0758530e4f7c req-9ddbb5ca-8564-4d65-9c71-1d6ec8110048 service nova] Acquiring lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.458928] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1288.459566] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.459738] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1288.460023] env[62814]: DEBUG oslo_concurrency.lockutils [req-8183e97a-431d-4e1d-8bba-0758530e4f7c req-9ddbb5ca-8564-4d65-9c71-1d6ec8110048 service nova] Acquired lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1288.460212] env[62814]: DEBUG nova.network.neutron [req-8183e97a-431d-4e1d-8bba-0758530e4f7c req-9ddbb5ca-8564-4d65-9c71-1d6ec8110048 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Refreshing network info cache for port d8936f3d-f722-43da-bf8e-17c76e3d95eb {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1288.461927] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8066eae8-7b18-4225-abe3-f2e2afa29b5f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.479630] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1288.479848] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1288.480023] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1288.480256] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1288.480457] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1288.480619] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1288.480829] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1288.480991] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1288.481179] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1288.481343] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1288.481517] env[62814]: DEBUG nova.virt.hardware [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1288.487773] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Reconfiguring VM to attach interface {{(pid=62814) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1288.488658] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2122c4a1-4c4d-4bd1-9433-0fddcf9cb777 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.506026] env[62814]: DEBUG oslo_vmware.api [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1288.506026] env[62814]: value = "task-4294853" [ 1288.506026] env[62814]: _type = "Task" [ 1288.506026] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.513689] env[62814]: DEBUG oslo_vmware.api [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294853, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.722089] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1289.015980] env[62814]: DEBUG oslo_vmware.api [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294853, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.168074] env[62814]: DEBUG nova.network.neutron [req-8183e97a-431d-4e1d-8bba-0758530e4f7c req-9ddbb5ca-8564-4d65-9c71-1d6ec8110048 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updated VIF entry in instance network info cache for port d8936f3d-f722-43da-bf8e-17c76e3d95eb. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1289.168525] env[62814]: DEBUG nova.network.neutron [req-8183e97a-431d-4e1d-8bba-0758530e4f7c req-9ddbb5ca-8564-4d65-9c71-1d6ec8110048 service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updating instance_info_cache with network_info: [{"id": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "address": "fa:16:3e:e7:24:4d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c053d0-8e", "ovs_interfaceid": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "d8936f3d-f722-43da-bf8e-17c76e3d95eb", "address": "fa:16:3e:c7:14:77", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8936f3d-f7", "ovs_interfaceid": "d8936f3d-f722-43da-bf8e-17c76e3d95eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1289.227372] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1289.227661] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.157s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1289.516503] env[62814]: DEBUG oslo_vmware.api [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294853, 'name': ReconfigVM_Task, 'duration_secs': 0.555634} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.516999] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1289.517236] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Reconfigured VM to attach interface {{(pid=62814) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1289.671285] env[62814]: DEBUG oslo_concurrency.lockutils [req-8183e97a-431d-4e1d-8bba-0758530e4f7c req-9ddbb5ca-8564-4d65-9c71-1d6ec8110048 service nova] Releasing lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1290.022174] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0ddd09e2-f590-4801-a299-482e07f174cf tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-7b82801d-7a0f-4baf-ad2e-566174d3eb35-d8936f3d-f722-43da-bf8e-17c76e3d95eb" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.201s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1291.227811] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.228164] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.228264] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.228421] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.502375] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1291.528747] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "interface-7b82801d-7a0f-4baf-ad2e-566174d3eb35-d8936f3d-f722-43da-bf8e-17c76e3d95eb" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1291.528969] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-7b82801d-7a0f-4baf-ad2e-566174d3eb35-d8936f3d-f722-43da-bf8e-17c76e3d95eb" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1292.031548] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1292.031733] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1292.032630] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2908b8f6-c8df-40e7-b5e6-7a751c473c21 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.050982] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94cae7f8-9c45-4475-aa21-6cf7b2289efc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.076786] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Reconfiguring VM to detach interface {{(pid=62814) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1292.077104] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1ac12e45-e62a-4ccc-b029-71a6a9b0405b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.096778] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1292.096778] env[62814]: value = "task-4294854" [ 1292.096778] env[62814]: _type = "Task" [ 1292.096778] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1292.106635] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.606416] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.106585] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1293.606822] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.107616] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1294.608722] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.108888] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1295.609403] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.109469] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.613117] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1296.931143] env[62814]: DEBUG oslo_concurrency.lockutils [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "d043b12d-a316-46ab-a30f-abec0a7963a6" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1296.931412] env[62814]: DEBUG oslo_concurrency.lockutils [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1296.931560] env[62814]: INFO nova.compute.manager [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Shelving [ 1297.110743] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.611417] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1297.940394] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1297.940783] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-409ec373-4702-407e-b572-3e99dcdf7701 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1297.948204] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1297.948204] env[62814]: value = "task-4294855" [ 1297.948204] env[62814]: _type = "Task" [ 1297.948204] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1297.956888] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294855, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1298.113019] env[62814]: DEBUG oslo_vmware.api [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294854, 'name': ReconfigVM_Task, 'duration_secs': 5.772613} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.113431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1298.113767] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Reconfigured VM to detach interface {{(pid=62814) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1298.458143] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294855, 'name': PowerOffVM_Task, 'duration_secs': 0.190951} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1298.458426] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1298.459220] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ef95a90-872f-40d6-b42c-1d63e1d59893 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.477235] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d5e0367-19bf-43ba-86f9-67e6b067d4ec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.987115] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1298.987510] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6db126aa-ae6b-4d55-a555-7a74c1bbf967 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1298.995414] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1298.995414] env[62814]: value = "task-4294856" [ 1298.995414] env[62814]: _type = "Task" [ 1298.995414] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1299.003605] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294856, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1299.394991] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1299.395265] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquired lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1299.395474] env[62814]: DEBUG nova.network.neutron [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1299.505628] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294856, 'name': CreateSnapshot_Task, 'duration_secs': 0.399133} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1299.505898] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1299.506653] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a88cef9-caba-45fd-b0ee-e8d0575c9a35 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1299.885934] env[62814]: DEBUG oslo_concurrency.lockutils [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1299.886249] env[62814]: DEBUG oslo_concurrency.lockutils [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1299.886467] env[62814]: DEBUG oslo_concurrency.lockutils [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1299.886649] env[62814]: DEBUG oslo_concurrency.lockutils [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1299.886816] env[62814]: DEBUG oslo_concurrency.lockutils [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1299.889031] env[62814]: INFO nova.compute.manager [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Terminating instance [ 1300.023538] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1300.023904] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-4d9d30d4-f9a7-4003-9dba-28ca2b8089da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.035408] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1300.035408] env[62814]: value = "task-4294857" [ 1300.035408] env[62814]: _type = "Task" [ 1300.035408] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.043873] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294857, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.137820] env[62814]: INFO nova.network.neutron [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Port d8936f3d-f722-43da-bf8e-17c76e3d95eb from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1300.138229] env[62814]: DEBUG nova.network.neutron [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updating instance_info_cache with network_info: [{"id": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "address": "fa:16:3e:e7:24:4d", "network": {"id": "1de03cb0-69fa-4200-8fa1-26375f44369a", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-139803101-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "10018060297c409d82935ad626cc54c9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe99da4f-5630-4afd-918b-b327193d8489", "external-id": "nsx-vlan-transportzone-688", "segmentation_id": 688, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90c053d0-8e", "ovs_interfaceid": "90c053d0-8e56-42b1-89da-b7acd45ac0fa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1300.396089] env[62814]: DEBUG nova.compute.manager [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1300.396378] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1300.397206] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15cad66-af7b-4f5c-a60f-81415cbdbe04 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.405018] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1300.405269] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e7d238a-7f4d-4570-bcb1-572b06165fb8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.411423] env[62814]: DEBUG oslo_vmware.api [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1300.411423] env[62814]: value = "task-4294858" [ 1300.411423] env[62814]: _type = "Task" [ 1300.411423] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1300.419562] env[62814]: DEBUG oslo_vmware.api [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294858, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.546323] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294857, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1300.641391] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Releasing lock "refresh_cache-7b82801d-7a0f-4baf-ad2e-566174d3eb35" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1300.921546] env[62814]: DEBUG oslo_vmware.api [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294858, 'name': PowerOffVM_Task, 'duration_secs': 0.16992} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1300.921867] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1300.922103] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1300.922393] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3bed6a63-4830-4120-86fd-6ae841485e36 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.986784] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1300.987024] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1300.987205] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleting the datastore file [datastore2] 7b82801d-7a0f-4baf-ad2e-566174d3eb35 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1300.987473] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5bc3597f-cdd5-4821-98f8-b5eb4167ea55 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1300.993910] env[62814]: DEBUG oslo_vmware.api [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1300.993910] env[62814]: value = "task-4294860" [ 1300.993910] env[62814]: _type = "Task" [ 1300.993910] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1301.001745] env[62814]: DEBUG oslo_vmware.api [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294860, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1301.046891] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294857, 'name': CloneVM_Task, 'duration_secs': 0.94507} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.047393] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Created linked-clone VM from snapshot [ 1301.047930] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f002a12-f2ca-42fd-8d58-1c0e0f3f698d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.055671] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Uploading image d6f2b60d-3670-4c14-b96f-fec9e2af6a24 {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1301.083191] env[62814]: DEBUG oslo_vmware.rw_handles [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1301.083191] env[62814]: value = "vm-845874" [ 1301.083191] env[62814]: _type = "VirtualMachine" [ 1301.083191] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1301.083517] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-0bfa1046-b6ce-4409-ae85-ca11de5a49ae {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.090310] env[62814]: DEBUG oslo_vmware.rw_handles [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lease: (returnval){ [ 1301.090310] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ccf17c-5bd7-fb82-1ca3-c6230f78607d" [ 1301.090310] env[62814]: _type = "HttpNfcLease" [ 1301.090310] env[62814]: } obtained for exporting VM: (result){ [ 1301.090310] env[62814]: value = "vm-845874" [ 1301.090310] env[62814]: _type = "VirtualMachine" [ 1301.090310] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1301.090633] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the lease: (returnval){ [ 1301.090633] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ccf17c-5bd7-fb82-1ca3-c6230f78607d" [ 1301.090633] env[62814]: _type = "HttpNfcLease" [ 1301.090633] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1301.096963] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1301.096963] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ccf17c-5bd7-fb82-1ca3-c6230f78607d" [ 1301.096963] env[62814]: _type = "HttpNfcLease" [ 1301.096963] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1301.145431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-72293027-b09f-4232-8421-552a4f8b6dde tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "interface-7b82801d-7a0f-4baf-ad2e-566174d3eb35-d8936f3d-f722-43da-bf8e-17c76e3d95eb" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.616s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1301.276204] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "add41cff-326f-46f4-b768-1f08ace2b31f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1301.276614] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "add41cff-326f-46f4-b768-1f08ace2b31f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1301.276902] env[62814]: DEBUG nova.compute.manager [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1301.278264] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4ce35f-54ec-40d0-bcb4-be5d24b4bb62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.286553] env[62814]: DEBUG nova.compute.manager [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1301.287169] env[62814]: DEBUG nova.objects.instance [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'flavor' on Instance uuid add41cff-326f-46f4-b768-1f08ace2b31f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1301.504257] env[62814]: DEBUG oslo_vmware.api [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294860, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164287} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1301.504521] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1301.504706] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1301.504886] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1301.505454] env[62814]: INFO nova.compute.manager [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1301.505454] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1301.505637] env[62814]: DEBUG nova.compute.manager [-] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1301.505637] env[62814]: DEBUG nova.network.neutron [-] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1301.602184] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1301.602184] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ccf17c-5bd7-fb82-1ca3-c6230f78607d" [ 1301.602184] env[62814]: _type = "HttpNfcLease" [ 1301.602184] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1301.602562] env[62814]: DEBUG oslo_vmware.rw_handles [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1301.602562] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ccf17c-5bd7-fb82-1ca3-c6230f78607d" [ 1301.602562] env[62814]: _type = "HttpNfcLease" [ 1301.602562] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1301.603575] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d32d8d3-4b3a-46ff-a151-b473a872c745 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1301.614556] env[62814]: DEBUG oslo_vmware.rw_handles [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527152af-bfc3-781b-5eaf-9175af059e4e/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1301.614880] env[62814]: DEBUG oslo_vmware.rw_handles [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527152af-bfc3-781b-5eaf-9175af059e4e/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1301.727786] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e1af389c-3e9e-43ae-a977-7e5e52abe50f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.295986] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1302.295986] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76227ea9-cc83-48da-8c1a-ba79115ef225 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.307344] env[62814]: DEBUG oslo_vmware.api [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1302.307344] env[62814]: value = "task-4294862" [ 1302.307344] env[62814]: _type = "Task" [ 1302.307344] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1302.317344] env[62814]: DEBUG oslo_vmware.api [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294862, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1302.346091] env[62814]: DEBUG nova.compute.manager [req-fc32d823-340b-4674-8790-f76adf4c545c req-8e7c0900-202f-4cf2-abd6-f8c2dedc005a service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Received event network-vif-deleted-90c053d0-8e56-42b1-89da-b7acd45ac0fa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1302.346091] env[62814]: INFO nova.compute.manager [req-fc32d823-340b-4674-8790-f76adf4c545c req-8e7c0900-202f-4cf2-abd6-f8c2dedc005a service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Neutron deleted interface 90c053d0-8e56-42b1-89da-b7acd45ac0fa; detaching it from the instance and deleting it from the info cache [ 1302.346584] env[62814]: DEBUG nova.network.neutron [req-fc32d823-340b-4674-8790-f76adf4c545c req-8e7c0900-202f-4cf2-abd6-f8c2dedc005a service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1302.817876] env[62814]: DEBUG oslo_vmware.api [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294862, 'name': PowerOffVM_Task, 'duration_secs': 0.326122} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1302.818201] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1302.818421] env[62814]: DEBUG nova.compute.manager [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1302.819377] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a5c923-5e8d-4573-86d3-2c32086e4236 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.823691] env[62814]: DEBUG nova.network.neutron [-] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1302.850948] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2ea7e9fe-d071-4ed5-bb01-834ccb41398e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.861674] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aaef08a-1ee8-4577-94fe-6dd6b5a6a48d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1302.892617] env[62814]: DEBUG nova.compute.manager [req-fc32d823-340b-4674-8790-f76adf4c545c req-8e7c0900-202f-4cf2-abd6-f8c2dedc005a service nova] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Detach interface failed, port_id=90c053d0-8e56-42b1-89da-b7acd45ac0fa, reason: Instance 7b82801d-7a0f-4baf-ad2e-566174d3eb35 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1303.330027] env[62814]: INFO nova.compute.manager [-] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Took 1.82 seconds to deallocate network for instance. [ 1303.334586] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ae0b87b-9b34-4946-8f23-e4086b77e1ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "add41cff-326f-46f4-b768-1f08ace2b31f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.058s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1303.791779] env[62814]: DEBUG nova.objects.instance [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'flavor' on Instance uuid add41cff-326f-46f4-b768-1f08ace2b31f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1303.840766] env[62814]: DEBUG oslo_concurrency.lockutils [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1303.841122] env[62814]: DEBUG oslo_concurrency.lockutils [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1303.841465] env[62814]: DEBUG nova.objects.instance [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'resources' on Instance uuid 7b82801d-7a0f-4baf-ad2e-566174d3eb35 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1304.297063] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1304.297291] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1304.297433] env[62814]: DEBUG nova.network.neutron [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1304.297645] env[62814]: DEBUG nova.objects.instance [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'info_cache' on Instance uuid add41cff-326f-46f4-b768-1f08ace2b31f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1304.424821] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee72e1ce-4500-48b9-935d-9d2afa73d7b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.434034] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a541e14-8023-4d1e-8485-7a4556bb6768 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.467017] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4290b544-1938-4a19-b0bf-ba363d89c1eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.474883] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3370f8e-a8c2-4401-81e0-985bbb0e989e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1304.490053] env[62814]: DEBUG nova.compute.provider_tree [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1304.801470] env[62814]: DEBUG nova.objects.base [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1304.993168] env[62814]: DEBUG nova.scheduler.client.report [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1305.498940] env[62814]: DEBUG oslo_concurrency.lockutils [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.657s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1305.529196] env[62814]: INFO nova.scheduler.client.report [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleted allocations for instance 7b82801d-7a0f-4baf-ad2e-566174d3eb35 [ 1305.543173] env[62814]: DEBUG nova.network.neutron [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Updating instance_info_cache with network_info: [{"id": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "address": "fa:16:3e:c3:d9:f1", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14c4e34d-9d", "ovs_interfaceid": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1306.037250] env[62814]: DEBUG oslo_concurrency.lockutils [None req-560ba5ed-eebb-4c70-87ef-0c5b78d89e7a tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "7b82801d-7a0f-4baf-ad2e-566174d3eb35" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.151s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1306.046109] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1306.567077] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1306.567546] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1306.567673] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "b5a2fb6c-7078-4f3c-b511-014beb96391d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1306.567769] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "b5a2fb6c-7078-4f3c-b511-014beb96391d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1306.567938] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "b5a2fb6c-7078-4f3c-b511-014beb96391d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1306.569863] env[62814]: INFO nova.compute.manager [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Terminating instance [ 1307.052038] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1307.052420] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27da3fe6-62cb-4511-8514-a0313781f509 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.062209] env[62814]: DEBUG oslo_vmware.api [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1307.062209] env[62814]: value = "task-4294863" [ 1307.062209] env[62814]: _type = "Task" [ 1307.062209] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.070298] env[62814]: DEBUG oslo_vmware.api [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.073018] env[62814]: DEBUG nova.compute.manager [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1307.073239] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1307.074034] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1fb3f1d-4b60-49ba-af99-1d98ff18d252 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.080991] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1307.081263] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8af7ee2c-3a24-4912-92bc-5a78f362331e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.087650] env[62814]: DEBUG oslo_vmware.api [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1307.087650] env[62814]: value = "task-4294864" [ 1307.087650] env[62814]: _type = "Task" [ 1307.087650] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.098535] env[62814]: DEBUG oslo_vmware.api [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294864, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1307.572745] env[62814]: DEBUG oslo_vmware.api [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294863, 'name': PowerOnVM_Task, 'duration_secs': 0.421068} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.573231] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1307.573338] env[62814]: DEBUG nova.compute.manager [None req-7514a15e-a79b-46e2-9c22-60797dab6c4e tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1307.574084] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c2f23c9-e41b-41c4-9856-7b6186d1225a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.596138] env[62814]: DEBUG oslo_vmware.api [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294864, 'name': PowerOffVM_Task, 'duration_secs': 0.245295} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1307.596415] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1307.596585] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1307.596871] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53e0bd76-fc83-40ee-83c8-2acee8dd4abc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.655385] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1307.655671] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1307.655785] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleting the datastore file [datastore2] b5a2fb6c-7078-4f3c-b511-014beb96391d {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1307.656086] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-68758a2a-7b06-45ae-9497-c3805259297e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1307.662948] env[62814]: DEBUG oslo_vmware.api [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for the task: (returnval){ [ 1307.662948] env[62814]: value = "task-4294866" [ 1307.662948] env[62814]: _type = "Task" [ 1307.662948] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1307.671163] env[62814]: DEBUG oslo_vmware.api [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294866, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.173293] env[62814]: DEBUG oslo_vmware.api [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Task: {'id': task-4294866, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190463} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1308.173639] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1308.173837] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1308.174110] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1308.174216] env[62814]: INFO nova.compute.manager [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1308.174452] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1308.174655] env[62814]: DEBUG nova.compute.manager [-] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1308.175599] env[62814]: DEBUG nova.network.neutron [-] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1308.511886] env[62814]: DEBUG nova.compute.manager [req-ab511c33-b9a6-408a-8f78-8b0598ec45e3 req-f573f551-d1c0-4d65-b1ee-e5216fdee5ca service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Received event network-vif-deleted-9edabfa8-9760-4ef4-adfb-afc6e2200b50 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1308.512195] env[62814]: INFO nova.compute.manager [req-ab511c33-b9a6-408a-8f78-8b0598ec45e3 req-f573f551-d1c0-4d65-b1ee-e5216fdee5ca service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Neutron deleted interface 9edabfa8-9760-4ef4-adfb-afc6e2200b50; detaching it from the instance and deleting it from the info cache [ 1308.512396] env[62814]: DEBUG nova.network.neutron [req-ab511c33-b9a6-408a-8f78-8b0598ec45e3 req-f573f551-d1c0-4d65-b1ee-e5216fdee5ca service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1308.763416] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6652c5-c70b-43a5-871e-9fb487cb27f9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.773467] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-205acd2f-ca8e-4970-ac7d-7dca2ec3ba23 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Suspending the VM {{(pid=62814) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1308.773971] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-80b35aa7-7de1-46c6-8e67-9817f77650e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1308.781591] env[62814]: DEBUG oslo_vmware.api [None req-205acd2f-ca8e-4970-ac7d-7dca2ec3ba23 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1308.781591] env[62814]: value = "task-4294867" [ 1308.781591] env[62814]: _type = "Task" [ 1308.781591] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1308.790940] env[62814]: DEBUG oslo_vmware.api [None req-205acd2f-ca8e-4970-ac7d-7dca2ec3ba23 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294867, 'name': SuspendVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1308.986350] env[62814]: DEBUG nova.network.neutron [-] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1309.016803] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cef9468e-d3ab-4ee3-86ae-d478d75ec5ef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.028101] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabd2cef-f04d-4072-a6f8-42c0eae80fc7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.056702] env[62814]: DEBUG nova.compute.manager [req-ab511c33-b9a6-408a-8f78-8b0598ec45e3 req-f573f551-d1c0-4d65-b1ee-e5216fdee5ca service nova] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Detach interface failed, port_id=9edabfa8-9760-4ef4-adfb-afc6e2200b50, reason: Instance b5a2fb6c-7078-4f3c-b511-014beb96391d could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1309.275125] env[62814]: DEBUG oslo_concurrency.lockutils [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1309.275470] env[62814]: DEBUG oslo_concurrency.lockutils [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1309.275708] env[62814]: INFO nova.compute.manager [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Shelving [ 1309.292368] env[62814]: DEBUG oslo_vmware.api [None req-205acd2f-ca8e-4970-ac7d-7dca2ec3ba23 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294867, 'name': SuspendVM_Task} progress is 54%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.489666] env[62814]: INFO nova.compute.manager [-] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Took 1.31 seconds to deallocate network for instance. [ 1309.792695] env[62814]: DEBUG oslo_vmware.api [None req-205acd2f-ca8e-4970-ac7d-7dca2ec3ba23 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294867, 'name': SuspendVM_Task, 'duration_secs': 0.727856} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1309.793105] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-205acd2f-ca8e-4970-ac7d-7dca2ec3ba23 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Suspended the VM {{(pid=62814) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1309.793838] env[62814]: DEBUG nova.compute.manager [None req-205acd2f-ca8e-4970-ac7d-7dca2ec3ba23 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1309.795405] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b59951-6e85-4929-a647-f4f30fb29ff9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.817452] env[62814]: DEBUG oslo_vmware.rw_handles [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527152af-bfc3-781b-5eaf-9175af059e4e/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1309.818565] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af4c8051-a077-48dc-a695-0993cd5b8413 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.824149] env[62814]: DEBUG oslo_vmware.rw_handles [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527152af-bfc3-781b-5eaf-9175af059e4e/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1309.824314] env[62814]: ERROR oslo_vmware.rw_handles [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527152af-bfc3-781b-5eaf-9175af059e4e/disk-0.vmdk due to incomplete transfer. [ 1309.824532] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-19c9dfe9-05dc-4cc9-bda3-480d174a584c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.831958] env[62814]: DEBUG oslo_vmware.rw_handles [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527152af-bfc3-781b-5eaf-9175af059e4e/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1309.832165] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Uploaded image d6f2b60d-3670-4c14-b96f-fec9e2af6a24 to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1309.834532] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1309.835055] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-94d7e259-e8ee-4670-9f47-8d1f06216f60 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1309.840663] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1309.840663] env[62814]: value = "task-4294868" [ 1309.840663] env[62814]: _type = "Task" [ 1309.840663] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1309.848901] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294868, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1309.996858] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1309.997227] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1309.997455] env[62814]: DEBUG nova.objects.instance [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lazy-loading 'resources' on Instance uuid b5a2fb6c-7078-4f3c-b511-014beb96391d {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1310.289471] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1310.289787] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1b9bfc98-f19c-422f-b778-6fe2afe15507 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.296941] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1310.296941] env[62814]: value = "task-4294869" [ 1310.296941] env[62814]: _type = "Task" [ 1310.296941] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.304250] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294869, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.350356] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294868, 'name': Destroy_Task, 'duration_secs': 0.340565} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1310.350658] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Destroyed the VM [ 1310.350934] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1310.351237] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d5bb4f06-a82e-40bb-a00d-fc478eb7ed8f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.357473] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1310.357473] env[62814]: value = "task-4294870" [ 1310.357473] env[62814]: _type = "Task" [ 1310.357473] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1310.367170] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294870, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1310.567437] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c40412c7-6190-4b28-98ac-b05a25893aed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.575347] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ae96b6-6564-4446-a631-0ccc942cb4ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.607344] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-474368ed-ea76-4e7f-b383-83c9501b8832 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.620157] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252b46a8-6459-4ee3-829f-04cec69815a4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.636486] env[62814]: DEBUG nova.compute.provider_tree [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1310.807288] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294869, 'name': PowerOffVM_Task, 'duration_secs': 0.283863} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1310.807665] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1310.808367] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c4c014d-ebaa-4525-bf8d-ddd229eb5a97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.827438] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7fa795-c555-4666-afa8-fb243cdd0f51 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1310.866510] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294870, 'name': RemoveSnapshot_Task, 'duration_secs': 0.372549} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1310.866760] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1310.867027] env[62814]: DEBUG nova.compute.manager [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1310.867743] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d549cd-7330-48e2-91ce-74d6651b7f90 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.119674] env[62814]: INFO nova.compute.manager [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Resuming [ 1311.120349] env[62814]: DEBUG nova.objects.instance [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'flavor' on Instance uuid add41cff-326f-46f4-b768-1f08ace2b31f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1311.139887] env[62814]: DEBUG nova.scheduler.client.report [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1311.337311] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Creating Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1311.337604] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f1776306-0bfc-46f6-a68a-a3627208ea3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.347562] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1311.347562] env[62814]: value = "task-4294871" [ 1311.347562] env[62814]: _type = "Task" [ 1311.347562] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.355258] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294871, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1311.378613] env[62814]: INFO nova.compute.manager [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Shelve offloading [ 1311.645067] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.648s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1311.672398] env[62814]: INFO nova.scheduler.client.report [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Deleted allocations for instance b5a2fb6c-7078-4f3c-b511-014beb96391d [ 1311.858024] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294871, 'name': CreateSnapshot_Task, 'duration_secs': 0.498717} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1311.858385] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Created Snapshot of the VM instance {{(pid=62814) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1311.859154] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe5a6b4-2890-4e36-9bc7-aa17cb47128e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.882800] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1311.883049] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e11920b1-15f4-4e6b-a239-d8898e49ca7c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.896303] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1311.896303] env[62814]: value = "task-4294872" [ 1311.896303] env[62814]: _type = "Task" [ 1311.896303] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1311.905596] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1311.905843] env[62814]: DEBUG nova.compute.manager [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1311.907315] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c7a9c2-f75e-45aa-8b5c-fd0d71315f44 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1311.913702] env[62814]: DEBUG oslo_concurrency.lockutils [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1311.913901] env[62814]: DEBUG oslo_concurrency.lockutils [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1311.914105] env[62814]: DEBUG nova.network.neutron [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1312.180593] env[62814]: DEBUG oslo_concurrency.lockutils [None req-fc5c548a-67ba-4b9e-a289-220f5d6a80ed tempest-AttachInterfacesTestJSON-1100165667 tempest-AttachInterfacesTestJSON-1100165667-project-member] Lock "b5a2fb6c-7078-4f3c-b511-014beb96391d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.613s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1312.376719] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Creating linked-clone VM from snapshot {{(pid=62814) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1312.377031] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9ddf6ceb-efb9-40c8-9930-008efa3e5ec5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.386405] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1312.386405] env[62814]: value = "task-4294873" [ 1312.386405] env[62814]: _type = "Task" [ 1312.386405] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1312.394262] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294873, 'name': CloneVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1312.630393] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1312.630657] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquired lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1312.630897] env[62814]: DEBUG nova.network.neutron [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1312.653865] env[62814]: DEBUG nova.network.neutron [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updating instance_info_cache with network_info: [{"id": "d584963d-9840-4607-b75d-b6a46b94cea3", "address": "fa:16:3e:ca:e5:53", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd584963d-98", "ovs_interfaceid": "d584963d-9840-4607-b75d-b6a46b94cea3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1312.898427] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294873, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.156387] env[62814]: DEBUG oslo_concurrency.lockutils [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1313.387912] env[62814]: DEBUG nova.network.neutron [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Updating instance_info_cache with network_info: [{"id": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "address": "fa:16:3e:c3:d9:f1", "network": {"id": "ab30c341-40f1-4c37-a993-88faca19d63b", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1973023050-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "513135674ee446d19fa8c667a47138a5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5b21ab10-d886-4453-9472-9e11fb3c450d", "external-id": "nsx-vlan-transportzone-885", "segmentation_id": 885, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap14c4e34d-9d", "ovs_interfaceid": "14c4e34d-9d53-41a8-ade9-86a6fa15cab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1313.403723] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294873, 'name': CloneVM_Task} progress is 94%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.454982] env[62814]: DEBUG nova.compute.manager [req-f181cfc5-3167-473e-90a0-86b12bff40c1 req-6c9c62bc-53ff-4d4c-be55-b18a45e43882 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received event network-vif-unplugged-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1313.456413] env[62814]: DEBUG oslo_concurrency.lockutils [req-f181cfc5-3167-473e-90a0-86b12bff40c1 req-6c9c62bc-53ff-4d4c-be55-b18a45e43882 service nova] Acquiring lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1313.456645] env[62814]: DEBUG oslo_concurrency.lockutils [req-f181cfc5-3167-473e-90a0-86b12bff40c1 req-6c9c62bc-53ff-4d4c-be55-b18a45e43882 service nova] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1313.456853] env[62814]: DEBUG oslo_concurrency.lockutils [req-f181cfc5-3167-473e-90a0-86b12bff40c1 req-6c9c62bc-53ff-4d4c-be55-b18a45e43882 service nova] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1313.457586] env[62814]: DEBUG nova.compute.manager [req-f181cfc5-3167-473e-90a0-86b12bff40c1 req-6c9c62bc-53ff-4d4c-be55-b18a45e43882 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] No waiting events found dispatching network-vif-unplugged-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1313.457826] env[62814]: WARNING nova.compute.manager [req-f181cfc5-3167-473e-90a0-86b12bff40c1 req-6c9c62bc-53ff-4d4c-be55-b18a45e43882 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received unexpected event network-vif-unplugged-d584963d-9840-4607-b75d-b6a46b94cea3 for instance with vm_state shelved and task_state shelving_offloading. [ 1313.644762] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1313.645957] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5926107-d07f-4d8f-b0eb-0d746013d4ea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.653677] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1313.653947] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d452ffc0-3017-4524-95c3-14566730a7eb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.816927] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1313.817172] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1313.817354] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleting the datastore file [datastore2] d043b12d-a316-46ab-a30f-abec0a7963a6 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1313.817619] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9a80cb3f-b481-4a36-8d6b-41a84a5e1624 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.823473] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1313.823473] env[62814]: value = "task-4294875" [ 1313.823473] env[62814]: _type = "Task" [ 1313.823473] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.831922] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294875, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.894951] env[62814]: DEBUG oslo_concurrency.lockutils [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Releasing lock "refresh_cache-add41cff-326f-46f4-b768-1f08ace2b31f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1313.899263] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a940a15f-d6d7-46b2-a4a6-9c69dd531ab3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.902257] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294873, 'name': CloneVM_Task, 'duration_secs': 1.363885} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1313.902535] env[62814]: INFO nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Created linked-clone VM from snapshot [ 1313.903701] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792e7f1e-c510-4f5a-8cab-a426ea9ca7ad {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.908028] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Resuming the VM {{(pid=62814) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1313.908574] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9dd76cb7-973c-4951-98a4-ee4ed94a40fb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.913675] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Uploading image 71021978-564a-44a6-9e72-98978e3943f2 {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1313.916878] env[62814]: DEBUG oslo_vmware.api [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1313.916878] env[62814]: value = "task-4294876" [ 1313.916878] env[62814]: _type = "Task" [ 1313.916878] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1313.929841] env[62814]: DEBUG oslo_vmware.api [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294876, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1313.944475] env[62814]: DEBUG oslo_vmware.rw_handles [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1313.944475] env[62814]: value = "vm-845876" [ 1313.944475] env[62814]: _type = "VirtualMachine" [ 1313.944475] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1313.944785] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-255a271c-ded1-4685-8c21-7a1cc6ecb7fc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.954289] env[62814]: DEBUG oslo_vmware.rw_handles [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lease: (returnval){ [ 1313.954289] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb115b-cb91-9cff-2b2b-2401f1c01f96" [ 1313.954289] env[62814]: _type = "HttpNfcLease" [ 1313.954289] env[62814]: } obtained for exporting VM: (result){ [ 1313.954289] env[62814]: value = "vm-845876" [ 1313.954289] env[62814]: _type = "VirtualMachine" [ 1313.954289] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1313.954647] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the lease: (returnval){ [ 1313.954647] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb115b-cb91-9cff-2b2b-2401f1c01f96" [ 1313.954647] env[62814]: _type = "HttpNfcLease" [ 1313.954647] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1313.962613] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1313.962613] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb115b-cb91-9cff-2b2b-2401f1c01f96" [ 1313.962613] env[62814]: _type = "HttpNfcLease" [ 1313.962613] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1314.333152] env[62814]: DEBUG oslo_vmware.api [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294875, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14877} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.333622] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1314.333622] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1314.333800] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1314.356844] env[62814]: INFO nova.scheduler.client.report [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleted allocations for instance d043b12d-a316-46ab-a30f-abec0a7963a6 [ 1314.429572] env[62814]: DEBUG oslo_vmware.api [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294876, 'name': PowerOnVM_Task} progress is 93%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1314.462683] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1314.462683] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb115b-cb91-9cff-2b2b-2401f1c01f96" [ 1314.462683] env[62814]: _type = "HttpNfcLease" [ 1314.462683] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1314.462983] env[62814]: DEBUG oslo_vmware.rw_handles [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1314.462983] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52cb115b-cb91-9cff-2b2b-2401f1c01f96" [ 1314.462983] env[62814]: _type = "HttpNfcLease" [ 1314.462983] env[62814]: }. {{(pid=62814) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1314.463761] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f628c9-9abe-4dcb-a8f3-1fa12993db15 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.471659] env[62814]: DEBUG oslo_vmware.rw_handles [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e3bda6-bab2-703d-2ab5-65a1b141ad93/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1314.471840] env[62814]: DEBUG oslo_vmware.rw_handles [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e3bda6-bab2-703d-2ab5-65a1b141ad93/disk-0.vmdk for reading. {{(pid=62814) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1314.656575] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-02ed289f-9a47-47a1-bf6a-c266838d9215 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1314.863431] env[62814]: DEBUG oslo_concurrency.lockutils [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1314.863604] env[62814]: DEBUG oslo_concurrency.lockutils [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1314.863830] env[62814]: DEBUG nova.objects.instance [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'resources' on Instance uuid d043b12d-a316-46ab-a30f-abec0a7963a6 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1314.930753] env[62814]: DEBUG oslo_vmware.api [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294876, 'name': PowerOnVM_Task, 'duration_secs': 0.597273} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1314.931223] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Resumed the VM {{(pid=62814) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1314.931223] env[62814]: DEBUG nova.compute.manager [None req-6b8e76e0-7750-44f0-9841-b884328d09ab tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1314.932132] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3499685a-6a98-46fb-92ea-2a3afdd7316b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.366857] env[62814]: DEBUG nova.objects.instance [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'numa_topology' on Instance uuid d043b12d-a316-46ab-a30f-abec0a7963a6 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1315.491125] env[62814]: DEBUG nova.compute.manager [req-96690ede-dcd9-4d71-ade3-87cd1cded4aa req-ec8ea129-c448-4cfa-8bce-fc006368eeef service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received event network-changed-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1315.493672] env[62814]: DEBUG nova.compute.manager [req-96690ede-dcd9-4d71-ade3-87cd1cded4aa req-ec8ea129-c448-4cfa-8bce-fc006368eeef service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Refreshing instance network info cache due to event network-changed-d584963d-9840-4607-b75d-b6a46b94cea3. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1315.495271] env[62814]: DEBUG oslo_concurrency.lockutils [req-96690ede-dcd9-4d71-ade3-87cd1cded4aa req-ec8ea129-c448-4cfa-8bce-fc006368eeef service nova] Acquiring lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1315.495271] env[62814]: DEBUG oslo_concurrency.lockutils [req-96690ede-dcd9-4d71-ade3-87cd1cded4aa req-ec8ea129-c448-4cfa-8bce-fc006368eeef service nova] Acquired lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1315.495271] env[62814]: DEBUG nova.network.neutron [req-96690ede-dcd9-4d71-ade3-87cd1cded4aa req-ec8ea129-c448-4cfa-8bce-fc006368eeef service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Refreshing network info cache for port d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1315.870739] env[62814]: DEBUG nova.objects.base [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1315.930781] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103521c5-1e77-4562-8bde-5bf4aa3e5132 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.940718] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98152649-130f-4433-a357-95e61b1769e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.971435] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67599738-37cf-43bd-bec0-456391da940f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.978981] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d2ff83c-3030-4ca9-8a7e-736e81b5a8b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1315.983092] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "add41cff-326f-46f4-b768-1f08ace2b31f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1315.983537] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "add41cff-326f-46f4-b768-1f08ace2b31f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1315.983537] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "add41cff-326f-46f4-b768-1f08ace2b31f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1315.984456] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "add41cff-326f-46f4-b768-1f08ace2b31f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1315.984456] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "add41cff-326f-46f4-b768-1f08ace2b31f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1315.986661] env[62814]: INFO nova.compute.manager [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Terminating instance [ 1315.995770] env[62814]: DEBUG nova.compute.provider_tree [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1316.237988] env[62814]: DEBUG nova.network.neutron [req-96690ede-dcd9-4d71-ade3-87cd1cded4aa req-ec8ea129-c448-4cfa-8bce-fc006368eeef service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updated VIF entry in instance network info cache for port d584963d-9840-4607-b75d-b6a46b94cea3. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1316.238508] env[62814]: DEBUG nova.network.neutron [req-96690ede-dcd9-4d71-ade3-87cd1cded4aa req-ec8ea129-c448-4cfa-8bce-fc006368eeef service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updating instance_info_cache with network_info: [{"id": "d584963d-9840-4607-b75d-b6a46b94cea3", "address": "fa:16:3e:ca:e5:53", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapd584963d-98", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1316.303296] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "d043b12d-a316-46ab-a30f-abec0a7963a6" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1316.500234] env[62814]: DEBUG nova.scheduler.client.report [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1316.504337] env[62814]: DEBUG nova.compute.manager [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1316.504585] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1316.505771] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-761cf62d-0d40-4cf5-8f4c-e6c1e6ed56a1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.513833] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1316.514131] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-461267dc-2519-48d3-acc9-5ff543a1f4fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1316.521697] env[62814]: DEBUG oslo_vmware.api [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1316.521697] env[62814]: value = "task-4294878" [ 1316.521697] env[62814]: _type = "Task" [ 1316.521697] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1316.530079] env[62814]: DEBUG oslo_vmware.api [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294878, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1316.741115] env[62814]: DEBUG oslo_concurrency.lockutils [req-96690ede-dcd9-4d71-ade3-87cd1cded4aa req-ec8ea129-c448-4cfa-8bce-fc006368eeef service nova] Releasing lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1317.007181] env[62814]: DEBUG oslo_concurrency.lockutils [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.143s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1317.031456] env[62814]: DEBUG oslo_vmware.api [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294878, 'name': PowerOffVM_Task, 'duration_secs': 0.199811} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.031794] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1317.031968] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1317.032219] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cbd5c9d9-40a5-4ea1-bd43-7d8f21017e0d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.091339] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1317.091528] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1317.091720] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleting the datastore file [datastore2] add41cff-326f-46f4-b768-1f08ace2b31f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1317.091982] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4de62ab1-7fa7-4263-bd32-8a1f19b78b3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1317.098405] env[62814]: DEBUG oslo_vmware.api [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for the task: (returnval){ [ 1317.098405] env[62814]: value = "task-4294880" [ 1317.098405] env[62814]: _type = "Task" [ 1317.098405] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1317.105953] env[62814]: DEBUG oslo_vmware.api [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294880, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1317.517333] env[62814]: DEBUG oslo_concurrency.lockutils [None req-af34e3fe-c655-4066-8395-6e056ad2d7ae tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 20.586s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1317.518399] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.215s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1317.518584] env[62814]: INFO nova.compute.manager [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Unshelving [ 1317.607925] env[62814]: DEBUG oslo_vmware.api [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Task: {'id': task-4294880, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164333} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1317.608208] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1317.608397] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1317.608574] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1317.608749] env[62814]: INFO nova.compute.manager [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1317.608993] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1317.609202] env[62814]: DEBUG nova.compute.manager [-] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1317.609297] env[62814]: DEBUG nova.network.neutron [-] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1318.101570] env[62814]: DEBUG nova.compute.manager [req-7852b1c6-6988-42dd-97d6-f6be960bcf49 req-5f4da53a-dac7-4c72-b751-59d0092c60ff service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Received event network-vif-deleted-14c4e34d-9d53-41a8-ade9-86a6fa15cab5 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1318.101897] env[62814]: INFO nova.compute.manager [req-7852b1c6-6988-42dd-97d6-f6be960bcf49 req-5f4da53a-dac7-4c72-b751-59d0092c60ff service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Neutron deleted interface 14c4e34d-9d53-41a8-ade9-86a6fa15cab5; detaching it from the instance and deleting it from the info cache [ 1318.102138] env[62814]: DEBUG nova.network.neutron [req-7852b1c6-6988-42dd-97d6-f6be960bcf49 req-5f4da53a-dac7-4c72-b751-59d0092c60ff service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1318.545293] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1318.545717] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1318.546093] env[62814]: DEBUG nova.objects.instance [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'pci_requests' on Instance uuid d043b12d-a316-46ab-a30f-abec0a7963a6 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1318.584952] env[62814]: DEBUG nova.network.neutron [-] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1318.604687] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed54809a-d49f-4813-84d1-3a61cb8166e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.616102] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccd85656-c481-44ea-9ea1-7e663092b0e6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1318.641671] env[62814]: DEBUG nova.compute.manager [req-7852b1c6-6988-42dd-97d6-f6be960bcf49 req-5f4da53a-dac7-4c72-b751-59d0092c60ff service nova] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Detach interface failed, port_id=14c4e34d-9d53-41a8-ade9-86a6fa15cab5, reason: Instance add41cff-326f-46f4-b768-1f08ace2b31f could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1319.049926] env[62814]: DEBUG nova.objects.instance [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'numa_topology' on Instance uuid d043b12d-a316-46ab-a30f-abec0a7963a6 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1319.088498] env[62814]: INFO nova.compute.manager [-] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Took 1.48 seconds to deallocate network for instance. [ 1319.552526] env[62814]: INFO nova.compute.claims [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1319.597366] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1320.620198] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38bec97c-0e97-4296-a188-a749a25f8492 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.628776] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fef31b2-b92b-4e17-87a0-9839ce26672a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.660098] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df440e10-3233-44fb-b41e-84f683336542 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.667654] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e127b0-2759-433a-a7c5-536ecab68856 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.681132] env[62814]: DEBUG nova.compute.provider_tree [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1321.184316] env[62814]: DEBUG nova.scheduler.client.report [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1321.690171] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.144s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1321.692449] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.095s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1321.692686] env[62814]: DEBUG nova.objects.instance [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lazy-loading 'resources' on Instance uuid add41cff-326f-46f4-b768-1f08ace2b31f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1321.722439] env[62814]: INFO nova.network.neutron [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updating port d584963d-9840-4607-b75d-b6a46b94cea3 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1322.247635] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7acb0502-b298-4602-9e35-2cb61c93209b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.255323] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf7fd67-f3c1-443f-986b-6bbdb35412f2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.285200] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d721902-eca3-44f8-af67-6c99c51ea3b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.292628] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4018af64-c1e3-436a-8bf8-9fdc22e9faa7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.305796] env[62814]: DEBUG nova.compute.provider_tree [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1322.575094] env[62814]: DEBUG oslo_vmware.rw_handles [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e3bda6-bab2-703d-2ab5-65a1b141ad93/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1322.576398] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2f3ce6-a6fd-4fbb-873f-384cb36aed2d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.583237] env[62814]: DEBUG oslo_vmware.rw_handles [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e3bda6-bab2-703d-2ab5-65a1b141ad93/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1322.583465] env[62814]: ERROR oslo_vmware.rw_handles [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e3bda6-bab2-703d-2ab5-65a1b141ad93/disk-0.vmdk due to incomplete transfer. [ 1322.583717] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-37714fd2-abdd-4f71-bd54-21914b092817 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.591074] env[62814]: DEBUG oslo_vmware.rw_handles [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e3bda6-bab2-703d-2ab5-65a1b141ad93/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1322.591310] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Uploaded image 71021978-564a-44a6-9e72-98978e3943f2 to the Glance image server {{(pid=62814) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1322.593747] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Destroying the VM {{(pid=62814) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1322.593988] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-03314415-8593-4888-96a9-e3b33d89ce70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1322.600275] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1322.600275] env[62814]: value = "task-4294881" [ 1322.600275] env[62814]: _type = "Task" [ 1322.600275] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1322.607994] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294881, 'name': Destroy_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1322.808944] env[62814]: DEBUG nova.scheduler.client.report [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1323.092433] env[62814]: DEBUG nova.compute.manager [req-eb5f97cc-c537-4ee6-b9bc-e203460f80f4 req-32062e20-ae6b-4e04-ba95-c3698aa75b88 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received event network-vif-plugged-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1323.092803] env[62814]: DEBUG oslo_concurrency.lockutils [req-eb5f97cc-c537-4ee6-b9bc-e203460f80f4 req-32062e20-ae6b-4e04-ba95-c3698aa75b88 service nova] Acquiring lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1323.092933] env[62814]: DEBUG oslo_concurrency.lockutils [req-eb5f97cc-c537-4ee6-b9bc-e203460f80f4 req-32062e20-ae6b-4e04-ba95-c3698aa75b88 service nova] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1323.093121] env[62814]: DEBUG oslo_concurrency.lockutils [req-eb5f97cc-c537-4ee6-b9bc-e203460f80f4 req-32062e20-ae6b-4e04-ba95-c3698aa75b88 service nova] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1323.093349] env[62814]: DEBUG nova.compute.manager [req-eb5f97cc-c537-4ee6-b9bc-e203460f80f4 req-32062e20-ae6b-4e04-ba95-c3698aa75b88 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] No waiting events found dispatching network-vif-plugged-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1323.093521] env[62814]: WARNING nova.compute.manager [req-eb5f97cc-c537-4ee6-b9bc-e203460f80f4 req-32062e20-ae6b-4e04-ba95-c3698aa75b88 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received unexpected event network-vif-plugged-d584963d-9840-4607-b75d-b6a46b94cea3 for instance with vm_state shelved_offloaded and task_state spawning. [ 1323.109301] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294881, 'name': Destroy_Task, 'duration_secs': 0.295302} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.109535] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Destroyed the VM [ 1323.110839] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Deleting Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1323.110839] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e2b0457a-45e2-4742-b17e-062261ac8d69 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.116040] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1323.116040] env[62814]: value = "task-4294882" [ 1323.116040] env[62814]: _type = "Task" [ 1323.116040] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1323.124486] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294882, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1323.189023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1323.189023] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1323.189023] env[62814]: DEBUG nova.network.neutron [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1323.314019] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.621s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1323.341448] env[62814]: INFO nova.scheduler.client.report [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Deleted allocations for instance add41cff-326f-46f4-b768-1f08ace2b31f [ 1323.625478] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294882, 'name': RemoveSnapshot_Task, 'duration_secs': 0.335527} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1323.625781] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Deleted Snapshot of the VM instance {{(pid=62814) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1323.626072] env[62814]: DEBUG nova.compute.manager [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1323.626851] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0e2ce1-eb6a-4956-93ca-959473e289e0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1323.848820] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ba29e5f9-50b1-48eb-9fde-7c14dc704414 tempest-ServerActionsTestJSON-2040483751 tempest-ServerActionsTestJSON-2040483751-project-member] Lock "add41cff-326f-46f4-b768-1f08ace2b31f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.865s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1323.913079] env[62814]: DEBUG nova.network.neutron [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updating instance_info_cache with network_info: [{"id": "d584963d-9840-4607-b75d-b6a46b94cea3", "address": "fa:16:3e:ca:e5:53", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd584963d-98", "ovs_interfaceid": "d584963d-9840-4607-b75d-b6a46b94cea3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1324.138885] env[62814]: INFO nova.compute.manager [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Shelve offloading [ 1324.415681] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1324.442331] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='dd72c3e5c19cc511e30d8303a131f82f',container_format='bare',created_at=2025-06-21T05:45:45Z,direct_url=,disk_format='vmdk',id=d6f2b60d-3670-4c14-b96f-fec9e2af6a24,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-717287423-shelved',owner='446e63a428c34d38ad07d01f0c90b23c',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2025-06-21T05:45:59Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1324.442601] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1324.442760] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1324.442940] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1324.443104] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1324.443251] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1324.443477] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1324.443688] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1324.443875] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1324.444050] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1324.444226] env[62814]: DEBUG nova.virt.hardware [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1324.445081] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2358a052-a409-40f4-90ef-d144d4d062bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.452930] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa5b7c7-553f-4a53-a6ca-b81096da1294 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.466868] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ca:e5:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b83383f-ed7a-4efd-aef7-aa8c15649d07', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd584963d-9840-4607-b75d-b6a46b94cea3', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1324.474357] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1324.474669] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1324.474896] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-96cf9b59-4727-49c0-8998-80cad93158f7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.493888] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1324.493888] env[62814]: value = "task-4294883" [ 1324.493888] env[62814]: _type = "Task" [ 1324.493888] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.501406] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294883, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1324.642974] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1324.643320] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61bbc245-0a96-4d8f-8903-17fd384914e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.651275] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1324.651275] env[62814]: value = "task-4294884" [ 1324.651275] env[62814]: _type = "Task" [ 1324.651275] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.663051] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1324.663278] env[62814]: DEBUG nova.compute.manager [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1324.664468] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0330728b-b3ef-4f66-8373-fa54ff7f0c22 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.670313] env[62814]: DEBUG oslo_concurrency.lockutils [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1324.670487] env[62814]: DEBUG oslo_concurrency.lockutils [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1324.670666] env[62814]: DEBUG nova.network.neutron [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1325.009008] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294883, 'name': CreateVM_Task, 'duration_secs': 0.276229} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.009304] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1325.009859] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1325.010063] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1325.010510] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1325.010842] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1bc4df9-1bf2-4758-9495-502ed006fc9f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.015892] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1325.015892] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dcc611-a5bc-9cd6-306c-cd1b4e53a376" [ 1325.015892] env[62814]: _type = "Task" [ 1325.015892] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.027549] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52dcc611-a5bc-9cd6-306c-cd1b4e53a376, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.117809] env[62814]: DEBUG nova.compute.manager [req-dd9b0d44-699d-4f97-98b5-992efaaabded req-2c92f589-3544-4279-99c7-50d7b6de1ad7 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received event network-changed-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1325.118033] env[62814]: DEBUG nova.compute.manager [req-dd9b0d44-699d-4f97-98b5-992efaaabded req-2c92f589-3544-4279-99c7-50d7b6de1ad7 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Refreshing instance network info cache due to event network-changed-d584963d-9840-4607-b75d-b6a46b94cea3. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1325.118255] env[62814]: DEBUG oslo_concurrency.lockutils [req-dd9b0d44-699d-4f97-98b5-992efaaabded req-2c92f589-3544-4279-99c7-50d7b6de1ad7 service nova] Acquiring lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1325.118392] env[62814]: DEBUG oslo_concurrency.lockutils [req-dd9b0d44-699d-4f97-98b5-992efaaabded req-2c92f589-3544-4279-99c7-50d7b6de1ad7 service nova] Acquired lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1325.118548] env[62814]: DEBUG nova.network.neutron [req-dd9b0d44-699d-4f97-98b5-992efaaabded req-2c92f589-3544-4279-99c7-50d7b6de1ad7 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Refreshing network info cache for port d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1325.528603] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1325.528983] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Processing image d6f2b60d-3670-4c14-b96f-fec9e2af6a24 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1325.529118] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24/d6f2b60d-3670-4c14-b96f-fec9e2af6a24.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1325.529267] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquired lock "[datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24/d6f2b60d-3670-4c14-b96f-fec9e2af6a24.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1325.529445] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1325.529696] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b48e5570-e06f-47c8-bca2-b082d92708e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.538375] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1325.538375] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1325.538975] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b129ddea-f5a5-44ce-bc12-de3048dc8ffe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.548814] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1325.548814] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52051b82-511e-b513-4220-15eacdcd24ae" [ 1325.548814] env[62814]: _type = "Task" [ 1325.548814] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.562986] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Preparing fetch location {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1325.563246] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Fetch image to [datastore2] OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc/OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc.vmdk {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1325.563473] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Downloading stream optimized image d6f2b60d-3670-4c14-b96f-fec9e2af6a24 to [datastore2] OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc/OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc.vmdk on the data store datastore2 as vApp {{(pid=62814) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1325.563666] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Downloading image file data d6f2b60d-3670-4c14-b96f-fec9e2af6a24 to the ESX as VM named 'OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc' {{(pid=62814) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1325.647114] env[62814]: DEBUG oslo_vmware.rw_handles [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1325.647114] env[62814]: value = "resgroup-9" [ 1325.647114] env[62814]: _type = "ResourcePool" [ 1325.647114] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1325.647434] env[62814]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-ccce95f3-1848-4e79-ad14-27196f704d44 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.662742] env[62814]: DEBUG nova.network.neutron [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updating instance_info_cache with network_info: [{"id": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "address": "fa:16:3e:02:56:99", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5888ff4-67", "ovs_interfaceid": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1325.669381] env[62814]: DEBUG oslo_vmware.rw_handles [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lease: (returnval){ [ 1325.669381] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527ecc70-cac3-e88d-d140-9a662ad05eb2" [ 1325.669381] env[62814]: _type = "HttpNfcLease" [ 1325.669381] env[62814]: } obtained for vApp import into resource pool (val){ [ 1325.669381] env[62814]: value = "resgroup-9" [ 1325.669381] env[62814]: _type = "ResourcePool" [ 1325.669381] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1325.669684] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the lease: (returnval){ [ 1325.669684] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527ecc70-cac3-e88d-d140-9a662ad05eb2" [ 1325.669684] env[62814]: _type = "HttpNfcLease" [ 1325.669684] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1325.676541] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1325.676541] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527ecc70-cac3-e88d-d140-9a662ad05eb2" [ 1325.676541] env[62814]: _type = "HttpNfcLease" [ 1325.676541] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1325.952312] env[62814]: DEBUG nova.network.neutron [req-dd9b0d44-699d-4f97-98b5-992efaaabded req-2c92f589-3544-4279-99c7-50d7b6de1ad7 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updated VIF entry in instance network info cache for port d584963d-9840-4607-b75d-b6a46b94cea3. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1325.952783] env[62814]: DEBUG nova.network.neutron [req-dd9b0d44-699d-4f97-98b5-992efaaabded req-2c92f589-3544-4279-99c7-50d7b6de1ad7 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updating instance_info_cache with network_info: [{"id": "d584963d-9840-4607-b75d-b6a46b94cea3", "address": "fa:16:3e:ca:e5:53", "network": {"id": "84c7202d-6585-4dcc-8a3b-1e013c63e820", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1786487996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "446e63a428c34d38ad07d01f0c90b23c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b83383f-ed7a-4efd-aef7-aa8c15649d07", "external-id": "nsx-vlan-transportzone-282", "segmentation_id": 282, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd584963d-98", "ovs_interfaceid": "d584963d-9840-4607-b75d-b6a46b94cea3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1326.165250] env[62814]: DEBUG oslo_concurrency.lockutils [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1326.177786] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1326.177786] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527ecc70-cac3-e88d-d140-9a662ad05eb2" [ 1326.177786] env[62814]: _type = "HttpNfcLease" [ 1326.177786] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1326.178103] env[62814]: DEBUG oslo_vmware.rw_handles [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1326.178103] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527ecc70-cac3-e88d-d140-9a662ad05eb2" [ 1326.178103] env[62814]: _type = "HttpNfcLease" [ 1326.178103] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1326.178817] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b17dc5-0065-4f10-bbab-2cecb2912d6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.185676] env[62814]: DEBUG oslo_vmware.rw_handles [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529e8555-dcd1-7f58-7567-547258118b04/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1326.185842] env[62814]: DEBUG oslo_vmware.rw_handles [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529e8555-dcd1-7f58-7567-547258118b04/disk-0.vmdk. {{(pid=62814) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1326.246334] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-76bde015-89ff-4c7b-9b71-bf8b9944209c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.444508] env[62814]: DEBUG nova.compute.manager [req-2816fa33-bffb-4613-9d56-236ea0469d26 req-11618130-2b8a-4110-a23a-8a8f0f884a2c service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received event network-vif-unplugged-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1326.444706] env[62814]: DEBUG oslo_concurrency.lockutils [req-2816fa33-bffb-4613-9d56-236ea0469d26 req-11618130-2b8a-4110-a23a-8a8f0f884a2c service nova] Acquiring lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1326.444921] env[62814]: DEBUG oslo_concurrency.lockutils [req-2816fa33-bffb-4613-9d56-236ea0469d26 req-11618130-2b8a-4110-a23a-8a8f0f884a2c service nova] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1326.445229] env[62814]: DEBUG oslo_concurrency.lockutils [req-2816fa33-bffb-4613-9d56-236ea0469d26 req-11618130-2b8a-4110-a23a-8a8f0f884a2c service nova] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1326.445622] env[62814]: DEBUG nova.compute.manager [req-2816fa33-bffb-4613-9d56-236ea0469d26 req-11618130-2b8a-4110-a23a-8a8f0f884a2c service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] No waiting events found dispatching network-vif-unplugged-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1326.445977] env[62814]: WARNING nova.compute.manager [req-2816fa33-bffb-4613-9d56-236ea0469d26 req-11618130-2b8a-4110-a23a-8a8f0f884a2c service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received unexpected event network-vif-unplugged-b5888ff4-67ff-4165-a2bc-390d621e4efa for instance with vm_state shelved and task_state shelving_offloading. [ 1326.455296] env[62814]: DEBUG oslo_concurrency.lockutils [req-dd9b0d44-699d-4f97-98b5-992efaaabded req-2c92f589-3544-4279-99c7-50d7b6de1ad7 service nova] Releasing lock "refresh_cache-d043b12d-a316-46ab-a30f-abec0a7963a6" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1326.542205] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1326.543133] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a34049-8c76-45e5-8c0b-1dd3da2f7089 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.553171] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1326.554909] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9eaddf81-88f3-4597-8c84-fd64ed95dc40 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.621027] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1326.621238] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1326.621562] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleting the datastore file [datastore2] ba933282-9711-4e99-ad2d-6ad4c6c516cc {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1326.626591] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2af9ad04-12e1-499a-a621-bd0b152533ef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1326.635986] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1326.635986] env[62814]: value = "task-4294887" [ 1326.635986] env[62814]: _type = "Task" [ 1326.635986] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1326.650358] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.146462] env[62814]: DEBUG oslo_vmware.api [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237972} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1327.146724] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1327.146916] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1327.147151] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1327.287201] env[62814]: INFO nova.scheduler.client.report [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleted allocations for instance ba933282-9711-4e99-ad2d-6ad4c6c516cc [ 1327.324056] env[62814]: DEBUG oslo_vmware.rw_handles [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Completed reading data from the image iterator. {{(pid=62814) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1327.325179] env[62814]: DEBUG oslo_vmware.rw_handles [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529e8555-dcd1-7f58-7567-547258118b04/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1327.326437] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba57e8e-e348-41e1-ae93-88f09eced182 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.333897] env[62814]: DEBUG oslo_vmware.rw_handles [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529e8555-dcd1-7f58-7567-547258118b04/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1327.334094] env[62814]: DEBUG oslo_vmware.rw_handles [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529e8555-dcd1-7f58-7567-547258118b04/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1327.334502] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-247adf89-9970-450d-a985-16d145176bc7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.575336] env[62814]: DEBUG oslo_vmware.rw_handles [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/529e8555-dcd1-7f58-7567-547258118b04/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1327.575637] env[62814]: INFO nova.virt.vmwareapi.images [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Downloaded image file data d6f2b60d-3670-4c14-b96f-fec9e2af6a24 [ 1327.576524] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa3e16d-9feb-4275-a60c-f5e9eb7961af {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.591914] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2b5e9d61-2491-48ee-9c22-bcdfdca75de1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.622609] env[62814]: INFO nova.virt.vmwareapi.images [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] The imported VM was unregistered [ 1327.626156] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Caching image {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1327.626156] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Creating directory with path [datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24 {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1327.626156] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4feb983-4af6-414e-9d4e-7df172939324 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.635833] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Created directory with path [datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24 {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1327.636027] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc/OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc.vmdk to [datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24/d6f2b60d-3670-4c14-b96f-fec9e2af6a24.vmdk. {{(pid=62814) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1327.636284] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-4f0360ac-fa7d-489f-9db6-ddb2e8f30019 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1327.641791] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1327.641791] env[62814]: value = "task-4294889" [ 1327.641791] env[62814]: _type = "Task" [ 1327.641791] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1327.649437] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294889, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.793202] env[62814]: DEBUG oslo_concurrency.lockutils [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1327.793535] env[62814]: DEBUG oslo_concurrency.lockutils [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1327.793790] env[62814]: DEBUG nova.objects.instance [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'resources' on Instance uuid ba933282-9711-4e99-ad2d-6ad4c6c516cc {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1328.152382] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294889, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.297209] env[62814]: DEBUG nova.objects.instance [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'numa_topology' on Instance uuid ba933282-9711-4e99-ad2d-6ad4c6c516cc {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1328.658922] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294889, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.800780] env[62814]: DEBUG nova.objects.base [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1328.811451] env[62814]: DEBUG nova.compute.manager [req-f68fef67-c02c-4fb1-b00b-bdbee0774d86 req-f6594849-900d-4690-8b4d-5284f6d7a4f9 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received event network-changed-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1328.811633] env[62814]: DEBUG nova.compute.manager [req-f68fef67-c02c-4fb1-b00b-bdbee0774d86 req-f6594849-900d-4690-8b4d-5284f6d7a4f9 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Refreshing instance network info cache due to event network-changed-b5888ff4-67ff-4165-a2bc-390d621e4efa. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1328.811844] env[62814]: DEBUG oslo_concurrency.lockutils [req-f68fef67-c02c-4fb1-b00b-bdbee0774d86 req-f6594849-900d-4690-8b4d-5284f6d7a4f9 service nova] Acquiring lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1328.811985] env[62814]: DEBUG oslo_concurrency.lockutils [req-f68fef67-c02c-4fb1-b00b-bdbee0774d86 req-f6594849-900d-4690-8b4d-5284f6d7a4f9 service nova] Acquired lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1328.812228] env[62814]: DEBUG nova.network.neutron [req-f68fef67-c02c-4fb1-b00b-bdbee0774d86 req-f6594849-900d-4690-8b4d-5284f6d7a4f9 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Refreshing network info cache for port b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1328.840276] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6cdaa1-ee72-40c8-9a0e-84cc34db60b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.848358] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a5770e-d410-4453-9572-fb4b476a875f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.881862] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f135244-30b5-44ce-940a-233f21a7f3d5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.890039] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be42046-c721-4453-af69-93bd65f11266 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.904434] env[62814]: DEBUG nova.compute.provider_tree [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1329.155020] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294889, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.411401] env[62814]: DEBUG nova.scheduler.client.report [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1329.638086] env[62814]: DEBUG nova.network.neutron [req-f68fef67-c02c-4fb1-b00b-bdbee0774d86 req-f6594849-900d-4690-8b4d-5284f6d7a4f9 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updated VIF entry in instance network info cache for port b5888ff4-67ff-4165-a2bc-390d621e4efa. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1329.638086] env[62814]: DEBUG nova.network.neutron [req-f68fef67-c02c-4fb1-b00b-bdbee0774d86 req-f6594849-900d-4690-8b4d-5284f6d7a4f9 service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updating instance_info_cache with network_info: [{"id": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "address": "fa:16:3e:02:56:99", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapb5888ff4-67", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1329.655890] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294889, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.918365] env[62814]: DEBUG oslo_concurrency.lockutils [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.124s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1330.139947] env[62814]: DEBUG oslo_concurrency.lockutils [req-f68fef67-c02c-4fb1-b00b-bdbee0774d86 req-f6594849-900d-4690-8b4d-5284f6d7a4f9 service nova] Releasing lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1330.155143] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294889, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.387977} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.155423] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc/OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc.vmdk to [datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24/d6f2b60d-3670-4c14-b96f-fec9e2af6a24.vmdk. [ 1330.155618] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Cleaning up location [datastore2] OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1330.155776] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_f4e45168-560d-4878-9df2-38ce10919bcc {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1330.156033] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ad701aa-9bd4-474c-8885-e852a467a94e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.161574] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1330.161574] env[62814]: value = "task-4294890" [ 1330.161574] env[62814]: _type = "Task" [ 1330.161574] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.169093] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294890, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.427119] env[62814]: DEBUG oslo_concurrency.lockutils [None req-956d3554-63d4-4e97-8c05-327ab88bc05d tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.152s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1330.449886] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1330.450155] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1330.450336] env[62814]: INFO nova.compute.manager [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Unshelving [ 1330.671626] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294890, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033461} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1330.671850] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1330.672030] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Releasing lock "[datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24/d6f2b60d-3670-4c14-b96f-fec9e2af6a24.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1330.672285] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24/d6f2b60d-3670-4c14-b96f-fec9e2af6a24.vmdk to [datastore2] d043b12d-a316-46ab-a30f-abec0a7963a6/d043b12d-a316-46ab-a30f-abec0a7963a6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1330.672567] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a5fdf7e1-3b8a-44c0-b9b6-bc997d0ca434 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1330.678972] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1330.678972] env[62814]: value = "task-4294891" [ 1330.678972] env[62814]: _type = "Task" [ 1330.678972] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1330.686958] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294891, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1330.887167] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "711f9623-4c61-47da-b7e5-15a41487b3b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1330.887167] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "711f9623-4c61-47da-b7e5-15a41487b3b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1331.189366] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294891, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.389820] env[62814]: DEBUG nova.compute.manager [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1331.477180] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1331.477604] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1331.477777] env[62814]: DEBUG nova.objects.instance [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'pci_requests' on Instance uuid ba933282-9711-4e99-ad2d-6ad4c6c516cc {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1331.690279] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294891, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1331.915101] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1331.981797] env[62814]: DEBUG nova.objects.instance [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'numa_topology' on Instance uuid ba933282-9711-4e99-ad2d-6ad4c6c516cc {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1332.190678] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294891, 'name': CopyVirtualDisk_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1332.484990] env[62814]: INFO nova.compute.claims [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1332.691095] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294891, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.191586] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294891, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.260703} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.191875] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/d6f2b60d-3670-4c14-b96f-fec9e2af6a24/d6f2b60d-3670-4c14-b96f-fec9e2af6a24.vmdk to [datastore2] d043b12d-a316-46ab-a30f-abec0a7963a6/d043b12d-a316-46ab-a30f-abec0a7963a6.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1333.192724] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28518fe8-4552-4a55-b295-fc2e1c84fb07 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.214931] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Reconfiguring VM instance instance-00000077 to attach disk [datastore2] d043b12d-a316-46ab-a30f-abec0a7963a6/d043b12d-a316-46ab-a30f-abec0a7963a6.vmdk or device None with type streamOptimized {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1333.215197] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26021132-9498-40d9-a19b-032aac3880dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.233970] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1333.233970] env[62814]: value = "task-4294892" [ 1333.233970] env[62814]: _type = "Task" [ 1333.233970] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.241610] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294892, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.541447] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9b45dc-073e-41af-97af-a0a3d29df0f4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.549341] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d611f11f-cda7-497b-b76c-0237e3a11bca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.578816] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8edaa93e-b5cb-4aca-bcb4-c78921abe365 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.585573] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84125fa9-ecf4-4b37-a491-2688dabfbaf8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.598225] env[62814]: DEBUG nova.compute.provider_tree [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1333.746125] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294892, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.101354] env[62814]: DEBUG nova.scheduler.client.report [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1334.244561] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294892, 'name': ReconfigVM_Task, 'duration_secs': 0.729327} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.245130] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Reconfigured VM instance instance-00000077 to attach disk [datastore2] d043b12d-a316-46ab-a30f-abec0a7963a6/d043b12d-a316-46ab-a30f-abec0a7963a6.vmdk or device None with type streamOptimized {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1334.245839] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6e3398e-4950-4256-a883-bc6bd33f8368 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.251366] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1334.251366] env[62814]: value = "task-4294893" [ 1334.251366] env[62814]: _type = "Task" [ 1334.251366] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.259166] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294893, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.606305] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.129s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1334.608701] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.694s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1334.610382] env[62814]: INFO nova.compute.claims [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1334.639789] env[62814]: INFO nova.network.neutron [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updating port b5888ff4-67ff-4165-a2bc-390d621e4efa with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1334.761128] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294893, 'name': Rename_Task, 'duration_secs': 0.137476} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.761476] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1334.761750] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bb40d75-b3bb-45bc-8d82-7bf84708a2e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.767982] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1334.767982] env[62814]: value = "task-4294894" [ 1334.767982] env[62814]: _type = "Task" [ 1334.767982] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.775064] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294894, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.277217] env[62814]: DEBUG oslo_vmware.api [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294894, 'name': PowerOnVM_Task, 'duration_secs': 0.414873} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1335.277479] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1335.376735] env[62814]: DEBUG nova.compute.manager [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1335.377672] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54a9bec-f55c-4637-804a-47ed8ef070e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.672546] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d48995-8a5c-417d-955e-5fe236870d3c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.680464] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2b5602-1ad7-4121-92ee-191b54083a8d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.711827] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1306747-147f-4984-abcf-c190e76671fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.718769] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94cc126b-f742-4616-893c-4c63ba2ce7e5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1335.731511] env[62814]: DEBUG nova.compute.provider_tree [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1335.895790] env[62814]: DEBUG oslo_concurrency.lockutils [None req-379a3e91-9332-46df-8148-253c6df445ed tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.377s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1336.032078] env[62814]: DEBUG nova.compute.manager [req-715e66ed-c25c-4b3d-83b8-5f27cc930c29 req-c1694f99-db39-469e-a744-c420e136f6af service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received event network-vif-plugged-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1336.032316] env[62814]: DEBUG oslo_concurrency.lockutils [req-715e66ed-c25c-4b3d-83b8-5f27cc930c29 req-c1694f99-db39-469e-a744-c420e136f6af service nova] Acquiring lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1336.032527] env[62814]: DEBUG oslo_concurrency.lockutils [req-715e66ed-c25c-4b3d-83b8-5f27cc930c29 req-c1694f99-db39-469e-a744-c420e136f6af service nova] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1336.032695] env[62814]: DEBUG oslo_concurrency.lockutils [req-715e66ed-c25c-4b3d-83b8-5f27cc930c29 req-c1694f99-db39-469e-a744-c420e136f6af service nova] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1336.032892] env[62814]: DEBUG nova.compute.manager [req-715e66ed-c25c-4b3d-83b8-5f27cc930c29 req-c1694f99-db39-469e-a744-c420e136f6af service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] No waiting events found dispatching network-vif-plugged-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1336.033085] env[62814]: WARNING nova.compute.manager [req-715e66ed-c25c-4b3d-83b8-5f27cc930c29 req-c1694f99-db39-469e-a744-c420e136f6af service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received unexpected event network-vif-plugged-b5888ff4-67ff-4165-a2bc-390d621e4efa for instance with vm_state shelved_offloaded and task_state spawning. [ 1336.116163] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1336.116349] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1336.116520] env[62814]: DEBUG nova.network.neutron [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1336.234672] env[62814]: DEBUG nova.scheduler.client.report [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1336.739303] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.131s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1336.739869] env[62814]: DEBUG nova.compute.manager [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1336.835673] env[62814]: DEBUG nova.network.neutron [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updating instance_info_cache with network_info: [{"id": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "address": "fa:16:3e:02:56:99", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5888ff4-67", "ovs_interfaceid": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1337.026801] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "d043b12d-a316-46ab-a30f-abec0a7963a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1337.027193] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1337.027475] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1337.027725] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1337.027958] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1337.030674] env[62814]: INFO nova.compute.manager [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Terminating instance [ 1337.246019] env[62814]: DEBUG nova.compute.utils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1337.247807] env[62814]: DEBUG nova.compute.manager [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1337.248855] env[62814]: DEBUG nova.network.neutron [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1337.296598] env[62814]: DEBUG nova.policy [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9516bbf4242140059eeb2182d3e7a58b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb7d25c2161f40db8197b1a56517d7f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1337.337946] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1337.365473] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='96ca82e964bac52fbdefcf84dda8c19b',container_format='bare',created_at=2025-06-21T05:45:58Z,direct_url=,disk_format='vmdk',id=71021978-564a-44a6-9e72-98978e3943f2,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-382934469-shelved',owner='8ed3e3bdfb4d4646a8c9eef1582cde85',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2025-06-21T05:46:11Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1337.365816] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1337.366010] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1337.366237] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1337.366418] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1337.366589] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1337.366786] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1337.366961] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1337.367139] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1337.367371] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1337.367502] env[62814]: DEBUG nova.virt.hardware [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1337.368440] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3155077e-9349-4ae9-8ab0-b2da7d65bd2e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.376223] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb1e9c7-f71f-42d8-a1c2-3977f2af993a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.390508] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:02:56:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b5888ff4-67ff-4165-a2bc-390d621e4efa', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1337.397729] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1337.397977] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1337.398200] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f512a170-0169-424e-8158-43f494471331 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.416900] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1337.416900] env[62814]: value = "task-4294895" [ 1337.416900] env[62814]: _type = "Task" [ 1337.416900] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.424182] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294895, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.535107] env[62814]: DEBUG nova.compute.manager [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1337.535377] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1337.536628] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f846045-44a3-4724-a9cd-50039a9bf68c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.544585] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1337.544821] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-302fc34e-4d5a-4930-b40b-b28c3e33c081 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.552448] env[62814]: DEBUG oslo_vmware.api [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1337.552448] env[62814]: value = "task-4294896" [ 1337.552448] env[62814]: _type = "Task" [ 1337.552448] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.559217] env[62814]: DEBUG oslo_vmware.api [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294896, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1337.613853] env[62814]: DEBUG nova.network.neutron [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Successfully created port: cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1337.751072] env[62814]: DEBUG nova.compute.manager [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1337.925496] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294895, 'name': CreateVM_Task, 'duration_secs': 0.309151} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1337.925678] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1337.926359] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1337.926525] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "[datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1337.926898] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1337.927164] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9a9fbd9-c8e3-4f8d-8c03-bc472bf3040d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1337.931232] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1337.931232] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529fddd3-bd05-5aff-80fe-955265dae593" [ 1337.931232] env[62814]: _type = "Task" [ 1337.931232] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1337.938313] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]529fddd3-bd05-5aff-80fe-955265dae593, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.059061] env[62814]: DEBUG nova.compute.manager [req-77c55168-e4dd-497e-b6fd-4eae1abdb90f req-403f8b66-347d-43df-9f7f-136127fe85bc service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received event network-changed-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1338.059277] env[62814]: DEBUG nova.compute.manager [req-77c55168-e4dd-497e-b6fd-4eae1abdb90f req-403f8b66-347d-43df-9f7f-136127fe85bc service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Refreshing instance network info cache due to event network-changed-b5888ff4-67ff-4165-a2bc-390d621e4efa. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1338.059769] env[62814]: DEBUG oslo_concurrency.lockutils [req-77c55168-e4dd-497e-b6fd-4eae1abdb90f req-403f8b66-347d-43df-9f7f-136127fe85bc service nova] Acquiring lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1338.059926] env[62814]: DEBUG oslo_concurrency.lockutils [req-77c55168-e4dd-497e-b6fd-4eae1abdb90f req-403f8b66-347d-43df-9f7f-136127fe85bc service nova] Acquired lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1338.060102] env[62814]: DEBUG nova.network.neutron [req-77c55168-e4dd-497e-b6fd-4eae1abdb90f req-403f8b66-347d-43df-9f7f-136127fe85bc service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Refreshing network info cache for port b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1338.064358] env[62814]: DEBUG oslo_vmware.api [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294896, 'name': PowerOffVM_Task, 'duration_secs': 0.212964} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.064840] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1338.065018] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1338.065512] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fe48135f-1a7e-48be-8a8d-f7746dd892e9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.124821] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1338.125048] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1338.125229] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleting the datastore file [datastore2] d043b12d-a316-46ab-a30f-abec0a7963a6 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1338.125480] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a572a833-6d71-4575-a8d9-8446522411f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.131410] env[62814]: DEBUG oslo_vmware.api [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for the task: (returnval){ [ 1338.131410] env[62814]: value = "task-4294898" [ 1338.131410] env[62814]: _type = "Task" [ 1338.131410] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.139296] env[62814]: DEBUG oslo_vmware.api [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294898, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.442914] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "[datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1338.443197] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Processing image 71021978-564a-44a6-9e72-98978e3943f2 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1338.443425] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2/71021978-564a-44a6-9e72-98978e3943f2.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1338.443569] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquired lock "[datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2/71021978-564a-44a6-9e72-98978e3943f2.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1338.443868] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1338.443983] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-81cda0be-248e-4e01-9c34-e2fd29b6b14e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.452424] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1338.452596] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1338.453284] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1ed5ee2-91ae-4556-a5a5-0582877218bd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.458197] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1338.458197] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e41f54-4dcc-0667-e837-249d4dbf39c1" [ 1338.458197] env[62814]: _type = "Task" [ 1338.458197] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1338.465202] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e41f54-4dcc-0667-e837-249d4dbf39c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1338.641719] env[62814]: DEBUG oslo_vmware.api [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Task: {'id': task-4294898, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13352} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1338.641999] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1338.642213] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1338.642398] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1338.642579] env[62814]: INFO nova.compute.manager [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1338.642835] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1338.643081] env[62814]: DEBUG nova.compute.manager [-] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1338.643155] env[62814]: DEBUG nova.network.neutron [-] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1338.760496] env[62814]: DEBUG nova.compute.manager [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1338.780828] env[62814]: DEBUG nova.network.neutron [req-77c55168-e4dd-497e-b6fd-4eae1abdb90f req-403f8b66-347d-43df-9f7f-136127fe85bc service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updated VIF entry in instance network info cache for port b5888ff4-67ff-4165-a2bc-390d621e4efa. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1338.781201] env[62814]: DEBUG nova.network.neutron [req-77c55168-e4dd-497e-b6fd-4eae1abdb90f req-403f8b66-347d-43df-9f7f-136127fe85bc service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updating instance_info_cache with network_info: [{"id": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "address": "fa:16:3e:02:56:99", "network": {"id": "b574d601-710c-4748-9b5a-9c2c4f4bf401", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1457908240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ed3e3bdfb4d4646a8c9eef1582cde85", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "022c7dd5-6c13-49c7-84f4-8b6c1fda4fb7", "external-id": "nsx-vlan-transportzone-694", "segmentation_id": 694, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb5888ff4-67", "ovs_interfaceid": "b5888ff4-67ff-4165-a2bc-390d621e4efa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1338.792209] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1338.792449] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1338.792607] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1338.792792] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1338.792933] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1338.793089] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1338.793542] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1338.793542] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1338.793694] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1338.793896] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1338.794239] env[62814]: DEBUG nova.virt.hardware [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1338.794943] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42caa021-544b-4510-9ff0-6595a4fc5175 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.805479] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4caa5c9-33f0-41b6-b40c-7706bce0d032 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.968940] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Preparing fetch location {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1338.969218] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Fetch image to [datastore2] OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7/OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7.vmdk {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1338.969451] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Downloading stream optimized image 71021978-564a-44a6-9e72-98978e3943f2 to [datastore2] OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7/OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7.vmdk on the data store datastore2 as vApp {{(pid=62814) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1338.969557] env[62814]: DEBUG nova.virt.vmwareapi.images [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Downloading image file data 71021978-564a-44a6-9e72-98978e3943f2 to the ESX as VM named 'OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7' {{(pid=62814) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1339.053639] env[62814]: DEBUG oslo_vmware.rw_handles [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1339.053639] env[62814]: value = "resgroup-9" [ 1339.053639] env[62814]: _type = "ResourcePool" [ 1339.053639] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1339.053984] env[62814]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-1f1d2497-4040-42b6-b36c-5e275674f0c0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.076305] env[62814]: DEBUG oslo_vmware.rw_handles [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lease: (returnval){ [ 1339.076305] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523e1d72-d0ca-e864-a5d8-be55d5511ad1" [ 1339.076305] env[62814]: _type = "HttpNfcLease" [ 1339.076305] env[62814]: } obtained for vApp import into resource pool (val){ [ 1339.076305] env[62814]: value = "resgroup-9" [ 1339.076305] env[62814]: _type = "ResourcePool" [ 1339.076305] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1339.076665] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the lease: (returnval){ [ 1339.076665] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523e1d72-d0ca-e864-a5d8-be55d5511ad1" [ 1339.076665] env[62814]: _type = "HttpNfcLease" [ 1339.076665] env[62814]: } to be ready. {{(pid=62814) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1339.084632] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1339.084632] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523e1d72-d0ca-e864-a5d8-be55d5511ad1" [ 1339.084632] env[62814]: _type = "HttpNfcLease" [ 1339.084632] env[62814]: } is initializing. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1339.264378] env[62814]: DEBUG nova.network.neutron [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Successfully updated port: cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1339.285821] env[62814]: DEBUG oslo_concurrency.lockutils [req-77c55168-e4dd-497e-b6fd-4eae1abdb90f req-403f8b66-347d-43df-9f7f-136127fe85bc service nova] Releasing lock "refresh_cache-ba933282-9711-4e99-ad2d-6ad4c6c516cc" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1339.585242] env[62814]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1339.585242] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523e1d72-d0ca-e864-a5d8-be55d5511ad1" [ 1339.585242] env[62814]: _type = "HttpNfcLease" [ 1339.585242] env[62814]: } is ready. {{(pid=62814) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1339.585538] env[62814]: DEBUG oslo_vmware.rw_handles [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1339.585538] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523e1d72-d0ca-e864-a5d8-be55d5511ad1" [ 1339.585538] env[62814]: _type = "HttpNfcLease" [ 1339.585538] env[62814]: }. {{(pid=62814) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1339.586343] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fdd7ba7-941f-4809-b829-228273c2974c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.594202] env[62814]: DEBUG oslo_vmware.rw_handles [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5261f6db-e905-6645-c965-ae8216f0b36a/disk-0.vmdk from lease info. {{(pid=62814) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1339.594383] env[62814]: DEBUG oslo_vmware.rw_handles [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5261f6db-e905-6645-c965-ae8216f0b36a/disk-0.vmdk. {{(pid=62814) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1339.654188] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-359c42dd-cd73-4820-a237-14c4e63b975f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.732433] env[62814]: DEBUG nova.network.neutron [-] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1339.766875] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1339.767131] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1339.767186] env[62814]: DEBUG nova.network.neutron [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1340.092080] env[62814]: DEBUG nova.compute.manager [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Received event network-vif-plugged-cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1340.092339] env[62814]: DEBUG oslo_concurrency.lockutils [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] Acquiring lock "711f9623-4c61-47da-b7e5-15a41487b3b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1340.092545] env[62814]: DEBUG oslo_concurrency.lockutils [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] Lock "711f9623-4c61-47da-b7e5-15a41487b3b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1340.092715] env[62814]: DEBUG oslo_concurrency.lockutils [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] Lock "711f9623-4c61-47da-b7e5-15a41487b3b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1340.092904] env[62814]: DEBUG nova.compute.manager [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] No waiting events found dispatching network-vif-plugged-cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1340.093089] env[62814]: WARNING nova.compute.manager [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Received unexpected event network-vif-plugged-cfcbae8d-a6f2-4583-9b46-101a3b73ac65 for instance with vm_state building and task_state spawning. [ 1340.093255] env[62814]: DEBUG nova.compute.manager [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Received event network-vif-deleted-d584963d-9840-4607-b75d-b6a46b94cea3 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1340.093423] env[62814]: DEBUG nova.compute.manager [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Received event network-changed-cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1340.093592] env[62814]: DEBUG nova.compute.manager [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Refreshing instance network info cache due to event network-changed-cfcbae8d-a6f2-4583-9b46-101a3b73ac65. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1340.093784] env[62814]: DEBUG oslo_concurrency.lockutils [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] Acquiring lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1340.237429] env[62814]: INFO nova.compute.manager [-] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Took 1.59 seconds to deallocate network for instance. [ 1340.314029] env[62814]: DEBUG nova.network.neutron [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1340.459261] env[62814]: DEBUG nova.network.neutron [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Updating instance_info_cache with network_info: [{"id": "cfcbae8d-a6f2-4583-9b46-101a3b73ac65", "address": "fa:16:3e:30:f5:07", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfcbae8d-a6", "ovs_interfaceid": "cfcbae8d-a6f2-4583-9b46-101a3b73ac65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1340.736825] env[62814]: DEBUG oslo_vmware.rw_handles [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Completed reading data from the image iterator. {{(pid=62814) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1340.737100] env[62814]: DEBUG oslo_vmware.rw_handles [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5261f6db-e905-6645-c965-ae8216f0b36a/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1340.738149] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7858f99f-7ad9-4a70-b6c4-628af3182ad3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.743912] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1340.744188] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1340.744408] env[62814]: DEBUG nova.objects.instance [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lazy-loading 'resources' on Instance uuid d043b12d-a316-46ab-a30f-abec0a7963a6 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1340.746939] env[62814]: DEBUG oslo_vmware.rw_handles [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5261f6db-e905-6645-c965-ae8216f0b36a/disk-0.vmdk is in state: ready. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1340.747109] env[62814]: DEBUG oslo_vmware.rw_handles [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5261f6db-e905-6645-c965-ae8216f0b36a/disk-0.vmdk. {{(pid=62814) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1340.747568] env[62814]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-6e954e3a-c85f-4d78-96d3-b304157b4b3d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.921869] env[62814]: DEBUG oslo_vmware.rw_handles [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5261f6db-e905-6645-c965-ae8216f0b36a/disk-0.vmdk. {{(pid=62814) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1340.922200] env[62814]: INFO nova.virt.vmwareapi.images [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Downloaded image file data 71021978-564a-44a6-9e72-98978e3943f2 [ 1340.922960] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37d34f62-84c6-45b4-8e66-ad4eee0271e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.938845] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc1a071b-5f86-41f1-8953-6fb52bce7eb6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.963267] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1340.963700] env[62814]: DEBUG nova.compute.manager [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Instance network_info: |[{"id": "cfcbae8d-a6f2-4583-9b46-101a3b73ac65", "address": "fa:16:3e:30:f5:07", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfcbae8d-a6", "ovs_interfaceid": "cfcbae8d-a6f2-4583-9b46-101a3b73ac65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1340.965095] env[62814]: DEBUG oslo_concurrency.lockutils [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] Acquired lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1340.965347] env[62814]: DEBUG nova.network.neutron [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Refreshing network info cache for port cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1340.966624] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:f5:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ffb921-796a-40fe-9662-d3fc01547dcb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cfcbae8d-a6f2-4583-9b46-101a3b73ac65', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1340.973882] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Creating folder: Project (eb7d25c2161f40db8197b1a56517d7f0). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1340.974165] env[62814]: INFO nova.virt.vmwareapi.images [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] The imported VM was unregistered [ 1340.976526] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Caching image {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1340.976789] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Creating directory with path [datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2 {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1340.977774] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58093ce1-3aa2-49d3-b749-3152d286e344 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.979436] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b71c19b1-8b85-4165-a1e8-8daee1a18506 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.990098] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Created folder: Project (eb7d25c2161f40db8197b1a56517d7f0) in parent group-v845547. [ 1340.990310] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Creating folder: Instances. Parent ref: group-v845881. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1340.991450] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-58662bd1-83b3-4858-9900-677b25a9ef97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1340.992990] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Created directory with path [datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2 {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1340.993191] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7/OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7.vmdk to [datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2/71021978-564a-44a6-9e72-98978e3943f2.vmdk. {{(pid=62814) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1340.993409] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-d87846de-aa05-4c51-95fd-089081e35d67 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.000036] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1341.000036] env[62814]: value = "task-4294903" [ 1341.000036] env[62814]: _type = "Task" [ 1341.000036] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.005062] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Created folder: Instances in parent group-v845881. [ 1341.005285] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1341.008152] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1341.008415] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294903, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.008613] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-99a2d5d0-eb49-40a9-a604-6e09a9fbd354 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.027040] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1341.027040] env[62814]: value = "task-4294904" [ 1341.027040] env[62814]: _type = "Task" [ 1341.027040] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1341.034070] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294904, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.311433] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac8652c-64aa-4375-a6a2-0d0094d9cbca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.319424] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ac87ab-8093-42af-8f06-10b7ed887d33 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.353727] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20045376-3131-46ec-813c-c2bce5a68ab1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.361733] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4d9356c-c2b5-4b56-92ec-a08adea3eb77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1341.375611] env[62814]: DEBUG nova.compute.provider_tree [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1341.513721] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294903, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.536965] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294904, 'name': CreateVM_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1341.683967] env[62814]: DEBUG nova.network.neutron [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Updated VIF entry in instance network info cache for port cfcbae8d-a6f2-4583-9b46-101a3b73ac65. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1341.684401] env[62814]: DEBUG nova.network.neutron [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Updating instance_info_cache with network_info: [{"id": "cfcbae8d-a6f2-4583-9b46-101a3b73ac65", "address": "fa:16:3e:30:f5:07", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfcbae8d-a6", "ovs_interfaceid": "cfcbae8d-a6f2-4583-9b46-101a3b73ac65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1341.879169] env[62814]: DEBUG nova.scheduler.client.report [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1342.010893] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294903, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.037535] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294904, 'name': CreateVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.187468] env[62814]: DEBUG oslo_concurrency.lockutils [req-bde9d2e4-a271-4645-b5bc-4f1abba15558 req-36cff321-0188-46aa-8b66-23eec02ff066 service nova] Releasing lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1342.385052] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1342.404279] env[62814]: INFO nova.scheduler.client.report [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Deleted allocations for instance d043b12d-a316-46ab-a30f-abec0a7963a6 [ 1342.511666] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294903, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.537245] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294904, 'name': CreateVM_Task, 'duration_secs': 1.22928} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1342.537419] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1342.538183] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1342.538352] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1342.538722] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1342.538964] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c20bfb6d-1ebb-468b-af16-7a127a569974 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1342.544049] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1342.544049] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aac1f4-97ee-43f8-4362-6596bf4726d3" [ 1342.544049] env[62814]: _type = "Task" [ 1342.544049] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1342.553385] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aac1f4-97ee-43f8-4362-6596bf4726d3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1342.912612] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d23c4dad-7753-411c-884d-a4ec0095c947 tempest-ServerActionsTestOtherB-634204386 tempest-ServerActionsTestOtherB-634204386-project-member] Lock "d043b12d-a316-46ab-a30f-abec0a7963a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.885s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1343.012788] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294903, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.054898] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52aac1f4-97ee-43f8-4362-6596bf4726d3, 'name': SearchDatastore_Task, 'duration_secs': 0.078082} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.055196] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1343.055449] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1343.055719] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1343.055874] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1343.056074] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1343.056348] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-57ae6312-8f07-426a-9235-38ce40576c5e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.073561] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1343.073852] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1343.074592] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30db6138-1d5b-49df-aa60-2eb47aa72bb1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.080101] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1343.080101] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f7cf1c-7cbf-bd69-6ebb-bb72777a8e41" [ 1343.080101] env[62814]: _type = "Task" [ 1343.080101] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1343.088076] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f7cf1c-7cbf-bd69-6ebb-bb72777a8e41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.513055] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294903, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.223494} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.513055] env[62814]: INFO nova.virt.vmwareapi.ds_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7/OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7.vmdk to [datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2/71021978-564a-44a6-9e72-98978e3943f2.vmdk. [ 1343.513055] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Cleaning up location [datastore2] OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1343.513055] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_cba9a6ff-6731-43d9-a87a-95e5af79c3d7 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1343.513412] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f43b341-7990-4f1b-9010-a176e3bc482b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.519712] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1343.519712] env[62814]: value = "task-4294906" [ 1343.519712] env[62814]: _type = "Task" [ 1343.519712] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1343.527123] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294906, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1343.589038] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52f7cf1c-7cbf-bd69-6ebb-bb72777a8e41, 'name': SearchDatastore_Task, 'duration_secs': 0.083032} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1343.589837] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9946f8c7-c9c2-4acc-90d1-0765ed27a705 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1343.594472] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1343.594472] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd909a-3c13-0611-08d1-c073a3f0050c" [ 1343.594472] env[62814]: _type = "Task" [ 1343.594472] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1343.601652] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd909a-3c13-0611-08d1-c073a3f0050c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.029495] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294906, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.036211} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.029762] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1344.029904] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Releasing lock "[datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2/71021978-564a-44a6-9e72-98978e3943f2.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1344.030171] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2/71021978-564a-44a6-9e72-98978e3943f2.vmdk to [datastore2] ba933282-9711-4e99-ad2d-6ad4c6c516cc/ba933282-9711-4e99-ad2d-6ad4c6c516cc.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1344.030411] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9e2eba73-f68f-4c7a-891c-d4993bc5f344 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.036816] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1344.036816] env[62814]: value = "task-4294907" [ 1344.036816] env[62814]: _type = "Task" [ 1344.036816] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.043870] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294907, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.103971] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52bd909a-3c13-0611-08d1-c073a3f0050c, 'name': SearchDatastore_Task, 'duration_secs': 0.008431} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.104199] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1344.104437] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 711f9623-4c61-47da-b7e5-15a41487b3b1/711f9623-4c61-47da-b7e5-15a41487b3b1.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1344.104680] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-46094476-2f63-4076-a2c6-135c478410e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.110447] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1344.110447] env[62814]: value = "task-4294908" [ 1344.110447] env[62814]: _type = "Task" [ 1344.110447] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.117834] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294908, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.546485] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294907, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.620116] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294908, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.047680] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294907, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.119630] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294908, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.549520] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294907, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.622161] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294908, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.501207} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.622603] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 711f9623-4c61-47da-b7e5-15a41487b3b1/711f9623-4c61-47da-b7e5-15a41487b3b1.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1345.622956] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1345.623330] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-088bd9fa-6613-4cd5-8c50-ef0853a278da {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.632032] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1345.632032] env[62814]: value = "task-4294909" [ 1345.632032] env[62814]: _type = "Task" [ 1345.632032] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.640821] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.050681] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294907, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.140727] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.32714} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.141018] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1346.141862] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6420dac-fdbc-41b4-8bf9-c9339c4a6033 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.164974] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Reconfiguring VM instance instance-0000007a to attach disk [datastore2] 711f9623-4c61-47da-b7e5-15a41487b3b1/711f9623-4c61-47da-b7e5-15a41487b3b1.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1346.165350] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b518805a-0de8-42a9-b2a5-2c90a5ec70b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.186078] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1346.186078] env[62814]: value = "task-4294910" [ 1346.186078] env[62814]: _type = "Task" [ 1346.186078] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.194767] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294910, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.507327] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1346.550380] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294907, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.370771} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.550881] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/71021978-564a-44a6-9e72-98978e3943f2/71021978-564a-44a6-9e72-98978e3943f2.vmdk to [datastore2] ba933282-9711-4e99-ad2d-6ad4c6c516cc/ba933282-9711-4e99-ad2d-6ad4c6c516cc.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1346.551650] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb17c989-19df-41de-ac2a-5cbd815dd9ee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.573099] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Reconfiguring VM instance instance-00000079 to attach disk [datastore2] ba933282-9711-4e99-ad2d-6ad4c6c516cc/ba933282-9711-4e99-ad2d-6ad4c6c516cc.vmdk or device None with type streamOptimized {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1346.573388] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4ac681a6-0042-4cfc-9ff1-e90678b28b27 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.594014] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1346.594014] env[62814]: value = "task-4294911" [ 1346.594014] env[62814]: _type = "Task" [ 1346.594014] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.603769] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294911, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.695757] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294910, 'name': ReconfigVM_Task, 'duration_secs': 0.497081} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.696109] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Reconfigured VM instance instance-0000007a to attach disk [datastore2] 711f9623-4c61-47da-b7e5-15a41487b3b1/711f9623-4c61-47da-b7e5-15a41487b3b1.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1346.696776] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-302f7532-cc38-4dfb-835f-bc40604a2764 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.703860] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1346.703860] env[62814]: value = "task-4294912" [ 1346.703860] env[62814]: _type = "Task" [ 1346.703860] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.711167] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294912, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.010280] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1347.010574] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1347.010745] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1347.010900] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1347.011853] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8523ad5-a6b3-454a-bf45-78a54d116d6d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.019717] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92126f48-f98e-445d-b48f-03b67dc419d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.034442] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec169988-cfb8-49dc-a570-2f29cb079485 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.040747] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f7ca18-5041-4299-9031-65203ea2147e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.069963] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180074MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1347.069963] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1347.069963] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1347.103526] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294911, 'name': ReconfigVM_Task, 'duration_secs': 0.282216} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.104595] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Reconfigured VM instance instance-00000079 to attach disk [datastore2] ba933282-9711-4e99-ad2d-6ad4c6c516cc/ba933282-9711-4e99-ad2d-6ad4c6c516cc.vmdk or device None with type streamOptimized {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1347.105244] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5780ea0-664a-43bd-b463-097508fa47fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.111921] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1347.111921] env[62814]: value = "task-4294913" [ 1347.111921] env[62814]: _type = "Task" [ 1347.111921] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.120092] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294913, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.213442] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294912, 'name': Rename_Task, 'duration_secs': 0.140152} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.213723] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1347.214013] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6dac86a5-e2aa-46f9-be6b-0e7f4b823f93 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.221529] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1347.221529] env[62814]: value = "task-4294914" [ 1347.221529] env[62814]: _type = "Task" [ 1347.221529] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.228984] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294914, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.622499] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294913, 'name': Rename_Task, 'duration_secs': 0.134359} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.622789] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1347.623067] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd9b5c8d-a4af-4029-afbe-c98379749255 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.629891] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1347.629891] env[62814]: value = "task-4294915" [ 1347.629891] env[62814]: _type = "Task" [ 1347.629891] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.638217] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294915, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.731527] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294914, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1348.094718] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance ba933282-9711-4e99-ad2d-6ad4c6c516cc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1348.094980] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 711f9623-4c61-47da-b7e5-15a41487b3b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1348.095077] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1348.095238] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=149GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1348.141261] env[62814]: DEBUG oslo_vmware.api [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294915, 'name': PowerOnVM_Task, 'duration_secs': 0.448882} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.142423] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1348.144966] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e9df04-73ca-4c05-8c1d-47ec0c32fdb4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.151407] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf4e015-54ee-458e-87ae-53db9912920d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.185178] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ce7547-2bc5-4dbc-b9ab-dc2751116f63 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.193079] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc35ce8-2566-4d1c-be74-ab414b60c36f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.206926] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1348.232626] env[62814]: DEBUG oslo_vmware.api [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294914, 'name': PowerOnVM_Task, 'duration_secs': 0.678517} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1348.233859] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1348.233859] env[62814]: INFO nova.compute.manager [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Took 9.47 seconds to spawn the instance on the hypervisor. [ 1348.233859] env[62814]: DEBUG nova.compute.manager [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1348.234108] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b17926-c5ea-4fa7-8917-e26835e37cea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.260984] env[62814]: DEBUG nova.compute.manager [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1348.262056] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751899ec-7f92-4590-a053-d92f8985b8cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1348.710117] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1348.751127] env[62814]: INFO nova.compute.manager [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Took 16.86 seconds to build instance. [ 1348.779880] env[62814]: DEBUG oslo_concurrency.lockutils [None req-b33ec383-ff8b-415b-9ed9-b4dd62a4c62c tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 18.329s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1349.215328] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1349.215697] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.146s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1349.252345] env[62814]: DEBUG oslo_concurrency.lockutils [None req-39288cc4-835c-4bc4-827d-26f297f34783 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "711f9623-4c61-47da-b7e5-15a41487b3b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.365s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1349.522187] env[62814]: DEBUG nova.compute.manager [req-6a15abf1-e5c7-4b69-b154-59d2aa33f2f1 req-96c69301-21c5-48a0-a05b-db56257d892c service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Received event network-changed-cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1349.523156] env[62814]: DEBUG nova.compute.manager [req-6a15abf1-e5c7-4b69-b154-59d2aa33f2f1 req-96c69301-21c5-48a0-a05b-db56257d892c service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Refreshing instance network info cache due to event network-changed-cfcbae8d-a6f2-4583-9b46-101a3b73ac65. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1349.523156] env[62814]: DEBUG oslo_concurrency.lockutils [req-6a15abf1-e5c7-4b69-b154-59d2aa33f2f1 req-96c69301-21c5-48a0-a05b-db56257d892c service nova] Acquiring lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1349.523156] env[62814]: DEBUG oslo_concurrency.lockutils [req-6a15abf1-e5c7-4b69-b154-59d2aa33f2f1 req-96c69301-21c5-48a0-a05b-db56257d892c service nova] Acquired lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1349.523156] env[62814]: DEBUG nova.network.neutron [req-6a15abf1-e5c7-4b69-b154-59d2aa33f2f1 req-96c69301-21c5-48a0-a05b-db56257d892c service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Refreshing network info cache for port cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1350.210727] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1350.210951] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1350.211126] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1350.211276] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1350.211433] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1350.211570] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 1350.246465] env[62814]: DEBUG nova.network.neutron [req-6a15abf1-e5c7-4b69-b154-59d2aa33f2f1 req-96c69301-21c5-48a0-a05b-db56257d892c service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Updated VIF entry in instance network info cache for port cfcbae8d-a6f2-4583-9b46-101a3b73ac65. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1350.246465] env[62814]: DEBUG nova.network.neutron [req-6a15abf1-e5c7-4b69-b154-59d2aa33f2f1 req-96c69301-21c5-48a0-a05b-db56257d892c service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Updating instance_info_cache with network_info: [{"id": "cfcbae8d-a6f2-4583-9b46-101a3b73ac65", "address": "fa:16:3e:30:f5:07", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfcbae8d-a6", "ovs_interfaceid": "cfcbae8d-a6f2-4583-9b46-101a3b73ac65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1350.748677] env[62814]: DEBUG oslo_concurrency.lockutils [req-6a15abf1-e5c7-4b69-b154-59d2aa33f2f1 req-96c69301-21c5-48a0-a05b-db56257d892c service nova] Releasing lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1351.349749] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquiring lock "54eb2a99-c7ef-4faf-a0d8-bef576081904" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1351.350070] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lock "54eb2a99-c7ef-4faf-a0d8-bef576081904" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1351.506731] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.507605] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1351.853721] env[62814]: DEBUG nova.compute.manager [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1352.376058] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1352.376355] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1352.377943] env[62814]: INFO nova.compute.claims [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1353.439946] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee24916-75d6-460e-bf94-3d310d96102e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.447903] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8a9c995-8e89-4743-8afd-0ff4973554ab {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.478808] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d88c86d-0683-4510-bd75-f368eea80465 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.485635] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b91144e3-33e3-4742-b113-e2819fc89968 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1353.498655] env[62814]: DEBUG nova.compute.provider_tree [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1354.001971] env[62814]: DEBUG nova.scheduler.client.report [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1354.506950] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.130s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1354.507528] env[62814]: DEBUG nova.compute.manager [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1355.012908] env[62814]: DEBUG nova.compute.utils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1355.014420] env[62814]: DEBUG nova.compute.manager [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1355.014589] env[62814]: DEBUG nova.network.neutron [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1355.071333] env[62814]: DEBUG nova.policy [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f611524d0c0c4fec8684265ffa9f57e3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e74fccd860b41ee94c3544dceb19ed0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1355.351507] env[62814]: DEBUG nova.network.neutron [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Successfully created port: cedb024a-64b9-483f-8aa5-1c160162053f {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1355.518182] env[62814]: DEBUG nova.compute.manager [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1356.528480] env[62814]: DEBUG nova.compute.manager [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1356.557597] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1356.557850] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1356.558017] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1356.558207] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1356.558357] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1356.558505] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1356.558713] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1356.558874] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1356.559051] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1356.559218] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1356.559389] env[62814]: DEBUG nova.virt.hardware [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1356.560295] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58408a87-431f-4239-8ddc-d6f9effaf1d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.568469] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1bea8e6-2347-4c2d-8c25-447368eae2b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.968158] env[62814]: DEBUG nova.compute.manager [req-0a949d66-507f-4909-b999-2f73c84bf68c req-372c482c-2cff-4984-8bb3-0e4b790f820a service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Received event network-vif-plugged-cedb024a-64b9-483f-8aa5-1c160162053f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1356.968395] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a949d66-507f-4909-b999-2f73c84bf68c req-372c482c-2cff-4984-8bb3-0e4b790f820a service nova] Acquiring lock "54eb2a99-c7ef-4faf-a0d8-bef576081904-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1356.968605] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a949d66-507f-4909-b999-2f73c84bf68c req-372c482c-2cff-4984-8bb3-0e4b790f820a service nova] Lock "54eb2a99-c7ef-4faf-a0d8-bef576081904-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1356.968769] env[62814]: DEBUG oslo_concurrency.lockutils [req-0a949d66-507f-4909-b999-2f73c84bf68c req-372c482c-2cff-4984-8bb3-0e4b790f820a service nova] Lock "54eb2a99-c7ef-4faf-a0d8-bef576081904-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1356.968935] env[62814]: DEBUG nova.compute.manager [req-0a949d66-507f-4909-b999-2f73c84bf68c req-372c482c-2cff-4984-8bb3-0e4b790f820a service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] No waiting events found dispatching network-vif-plugged-cedb024a-64b9-483f-8aa5-1c160162053f {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1356.969116] env[62814]: WARNING nova.compute.manager [req-0a949d66-507f-4909-b999-2f73c84bf68c req-372c482c-2cff-4984-8bb3-0e4b790f820a service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Received unexpected event network-vif-plugged-cedb024a-64b9-483f-8aa5-1c160162053f for instance with vm_state building and task_state spawning. [ 1357.052016] env[62814]: DEBUG nova.network.neutron [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Successfully updated port: cedb024a-64b9-483f-8aa5-1c160162053f {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1357.555469] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquiring lock "refresh_cache-54eb2a99-c7ef-4faf-a0d8-bef576081904" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1357.555738] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquired lock "refresh_cache-54eb2a99-c7ef-4faf-a0d8-bef576081904" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1357.555776] env[62814]: DEBUG nova.network.neutron [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1358.087160] env[62814]: DEBUG nova.network.neutron [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1358.201217] env[62814]: DEBUG nova.network.neutron [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Updating instance_info_cache with network_info: [{"id": "cedb024a-64b9-483f-8aa5-1c160162053f", "address": "fa:16:3e:23:36:c7", "network": {"id": "1b0dd8be-cd7f-442d-8374-9f548bf2a40b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-245217674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e74fccd860b41ee94c3544dceb19ed0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcedb024a-64", "ovs_interfaceid": "cedb024a-64b9-483f-8aa5-1c160162053f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1358.704210] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Releasing lock "refresh_cache-54eb2a99-c7ef-4faf-a0d8-bef576081904" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1358.704584] env[62814]: DEBUG nova.compute.manager [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Instance network_info: |[{"id": "cedb024a-64b9-483f-8aa5-1c160162053f", "address": "fa:16:3e:23:36:c7", "network": {"id": "1b0dd8be-cd7f-442d-8374-9f548bf2a40b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-245217674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e74fccd860b41ee94c3544dceb19ed0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcedb024a-64", "ovs_interfaceid": "cedb024a-64b9-483f-8aa5-1c160162053f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1358.704947] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:36:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68ec9c06-8680-4a41-abad-cddbd1f768c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cedb024a-64b9-483f-8aa5-1c160162053f', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1358.712541] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Creating folder: Project (6e74fccd860b41ee94c3544dceb19ed0). Parent ref: group-v845547. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1358.712805] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b6d5269-5edf-4835-9d13-5e0cfe2e62c1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.725521] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Created folder: Project (6e74fccd860b41ee94c3544dceb19ed0) in parent group-v845547. [ 1358.725696] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Creating folder: Instances. Parent ref: group-v845884. {{(pid=62814) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1358.725950] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6df1a900-0de7-44a7-9bbc-7f4c471e7502 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.734810] env[62814]: INFO nova.virt.vmwareapi.vm_util [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Created folder: Instances in parent group-v845884. [ 1358.735033] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1358.735214] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1358.735395] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1362abb-37aa-417a-8f52-b7cf797e5be8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1358.752065] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1358.752065] env[62814]: value = "task-4294918" [ 1358.752065] env[62814]: _type = "Task" [ 1358.752065] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1358.759091] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294918, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1358.996234] env[62814]: DEBUG nova.compute.manager [req-7d5e3631-4312-46cb-969d-75def1601c3a req-dc56f9e9-0606-4d3f-83bb-20bcc8ce18b0 service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Received event network-changed-cedb024a-64b9-483f-8aa5-1c160162053f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1358.996234] env[62814]: DEBUG nova.compute.manager [req-7d5e3631-4312-46cb-969d-75def1601c3a req-dc56f9e9-0606-4d3f-83bb-20bcc8ce18b0 service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Refreshing instance network info cache due to event network-changed-cedb024a-64b9-483f-8aa5-1c160162053f. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1358.996234] env[62814]: DEBUG oslo_concurrency.lockutils [req-7d5e3631-4312-46cb-969d-75def1601c3a req-dc56f9e9-0606-4d3f-83bb-20bcc8ce18b0 service nova] Acquiring lock "refresh_cache-54eb2a99-c7ef-4faf-a0d8-bef576081904" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1358.996234] env[62814]: DEBUG oslo_concurrency.lockutils [req-7d5e3631-4312-46cb-969d-75def1601c3a req-dc56f9e9-0606-4d3f-83bb-20bcc8ce18b0 service nova] Acquired lock "refresh_cache-54eb2a99-c7ef-4faf-a0d8-bef576081904" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1358.996234] env[62814]: DEBUG nova.network.neutron [req-7d5e3631-4312-46cb-969d-75def1601c3a req-dc56f9e9-0606-4d3f-83bb-20bcc8ce18b0 service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Refreshing network info cache for port cedb024a-64b9-483f-8aa5-1c160162053f {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1359.262141] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294918, 'name': CreateVM_Task, 'duration_secs': 0.280739} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.262479] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1359.263382] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1359.263556] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1359.263881] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1359.264235] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c73128a-f8ba-4b94-a315-6fda7de56084 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.268976] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for the task: (returnval){ [ 1359.268976] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527b09c6-e2af-568f-0e7b-5afd972db853" [ 1359.268976] env[62814]: _type = "Task" [ 1359.268976] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.275929] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527b09c6-e2af-568f-0e7b-5afd972db853, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1359.665700] env[62814]: DEBUG nova.network.neutron [req-7d5e3631-4312-46cb-969d-75def1601c3a req-dc56f9e9-0606-4d3f-83bb-20bcc8ce18b0 service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Updated VIF entry in instance network info cache for port cedb024a-64b9-483f-8aa5-1c160162053f. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1359.666109] env[62814]: DEBUG nova.network.neutron [req-7d5e3631-4312-46cb-969d-75def1601c3a req-dc56f9e9-0606-4d3f-83bb-20bcc8ce18b0 service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Updating instance_info_cache with network_info: [{"id": "cedb024a-64b9-483f-8aa5-1c160162053f", "address": "fa:16:3e:23:36:c7", "network": {"id": "1b0dd8be-cd7f-442d-8374-9f548bf2a40b", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-245217674-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e74fccd860b41ee94c3544dceb19ed0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcedb024a-64", "ovs_interfaceid": "cedb024a-64b9-483f-8aa5-1c160162053f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1359.779210] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527b09c6-e2af-568f-0e7b-5afd972db853, 'name': SearchDatastore_Task, 'duration_secs': 0.009482} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1359.779537] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1359.779678] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1359.779912] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1359.780075] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1359.780259] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1359.780507] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1f341da-f00b-42c2-9148-659526ac5cf0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.788761] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1359.788922] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1359.789606] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29d52238-cb2c-41b7-9f2c-7a234d906d62 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.794531] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for the task: (returnval){ [ 1359.794531] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e4d7c7-164d-6630-d90a-42f99a4be8f8" [ 1359.794531] env[62814]: _type = "Task" [ 1359.794531] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1359.801556] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e4d7c7-164d-6630-d90a-42f99a4be8f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.168396] env[62814]: DEBUG oslo_concurrency.lockutils [req-7d5e3631-4312-46cb-969d-75def1601c3a req-dc56f9e9-0606-4d3f-83bb-20bcc8ce18b0 service nova] Releasing lock "refresh_cache-54eb2a99-c7ef-4faf-a0d8-bef576081904" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1360.305933] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e4d7c7-164d-6630-d90a-42f99a4be8f8, 'name': SearchDatastore_Task, 'duration_secs': 0.01122} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.306690] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b687bd26-34b0-4a36-b7bc-50d1a212d456 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.312081] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for the task: (returnval){ [ 1360.312081] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c6017a-f984-3d76-e37c-689f18df46e0" [ 1360.312081] env[62814]: _type = "Task" [ 1360.312081] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.319123] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c6017a-f984-3d76-e37c-689f18df46e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.823048] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c6017a-f984-3d76-e37c-689f18df46e0, 'name': SearchDatastore_Task, 'duration_secs': 0.009963} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.823048] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1360.823377] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 54eb2a99-c7ef-4faf-a0d8-bef576081904/54eb2a99-c7ef-4faf-a0d8-bef576081904.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1360.823473] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f6ffc999-eb71-4a70-a312-ae8fba70e317 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.829332] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for the task: (returnval){ [ 1360.829332] env[62814]: value = "task-4294919" [ 1360.829332] env[62814]: _type = "Task" [ 1360.829332] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.836414] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294919, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.339289] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294919, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48619} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.339599] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 54eb2a99-c7ef-4faf-a0d8-bef576081904/54eb2a99-c7ef-4faf-a0d8-bef576081904.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1361.339856] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1361.340162] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f356c5ad-f8cb-4138-9af5-38d694e5f57e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.347839] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for the task: (returnval){ [ 1361.347839] env[62814]: value = "task-4294920" [ 1361.347839] env[62814]: _type = "Task" [ 1361.347839] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.356532] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294920, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.858360] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294920, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06732} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1361.858360] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1361.859137] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9640842e-f6a5-4be3-b1e4-fef0d2ae185b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.881222] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Reconfiguring VM instance instance-0000007b to attach disk [datastore2] 54eb2a99-c7ef-4faf-a0d8-bef576081904/54eb2a99-c7ef-4faf-a0d8-bef576081904.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1361.881596] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff1f56c7-2211-4fd9-9fec-cb66a3c11144 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.901588] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for the task: (returnval){ [ 1361.901588] env[62814]: value = "task-4294921" [ 1361.901588] env[62814]: _type = "Task" [ 1361.901588] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.909347] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294921, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.411360] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294921, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.911580] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294921, 'name': ReconfigVM_Task, 'duration_secs': 0.928589} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.911922] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Reconfigured VM instance instance-0000007b to attach disk [datastore2] 54eb2a99-c7ef-4faf-a0d8-bef576081904/54eb2a99-c7ef-4faf-a0d8-bef576081904.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1362.912565] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3411b35-f0c4-4d29-b6b7-c259b2b850e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.918926] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for the task: (returnval){ [ 1362.918926] env[62814]: value = "task-4294922" [ 1362.918926] env[62814]: _type = "Task" [ 1362.918926] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.926038] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294922, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.429643] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294922, 'name': Rename_Task, 'duration_secs': 0.143631} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.429943] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1363.430208] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-489fc38d-33ed-40af-aae8-35d3cdadd35f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.436311] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for the task: (returnval){ [ 1363.436311] env[62814]: value = "task-4294923" [ 1363.436311] env[62814]: _type = "Task" [ 1363.436311] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.443632] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294923, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.946616] env[62814]: DEBUG oslo_vmware.api [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294923, 'name': PowerOnVM_Task, 'duration_secs': 0.441069} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.946950] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1363.947200] env[62814]: INFO nova.compute.manager [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Took 7.42 seconds to spawn the instance on the hypervisor. [ 1363.947415] env[62814]: DEBUG nova.compute.manager [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1363.948245] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23ac875-60fc-47a7-9aa6-21094d76e507 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.465983] env[62814]: INFO nova.compute.manager [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Took 12.11 seconds to build instance. [ 1364.968631] env[62814]: DEBUG oslo_concurrency.lockutils [None req-dfe2c377-4d3e-4046-b078-3b2311918907 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lock "54eb2a99-c7ef-4faf-a0d8-bef576081904" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.618s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1366.285147] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquiring lock "54eb2a99-c7ef-4faf-a0d8-bef576081904" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1366.285623] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lock "54eb2a99-c7ef-4faf-a0d8-bef576081904" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1366.285847] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquiring lock "54eb2a99-c7ef-4faf-a0d8-bef576081904-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1366.286151] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lock "54eb2a99-c7ef-4faf-a0d8-bef576081904-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1366.286389] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lock "54eb2a99-c7ef-4faf-a0d8-bef576081904-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1366.288555] env[62814]: INFO nova.compute.manager [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Terminating instance [ 1366.792717] env[62814]: DEBUG nova.compute.manager [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1366.792997] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1366.794024] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db5f410-4ea5-4b89-939e-42a5c8b022df {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.802027] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1366.802254] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cb1f30c-d380-4cfe-880f-2f7c4cebe7de {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.809106] env[62814]: DEBUG oslo_vmware.api [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for the task: (returnval){ [ 1366.809106] env[62814]: value = "task-4294924" [ 1366.809106] env[62814]: _type = "Task" [ 1366.809106] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.817835] env[62814]: DEBUG oslo_vmware.api [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294924, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.318709] env[62814]: DEBUG oslo_vmware.api [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294924, 'name': PowerOffVM_Task, 'duration_secs': 0.204014} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.319070] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1367.319111] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1367.319340] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-905e84eb-f1ec-4fb1-a5f6-b93dbb93d42e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.472837] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1367.473086] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1367.473300] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Deleting the datastore file [datastore2] 54eb2a99-c7ef-4faf-a0d8-bef576081904 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1367.473657] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1ab9d0a-5984-432e-91ac-d2e3f770aa66 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1367.479877] env[62814]: DEBUG oslo_vmware.api [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for the task: (returnval){ [ 1367.479877] env[62814]: value = "task-4294926" [ 1367.479877] env[62814]: _type = "Task" [ 1367.479877] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1367.487856] env[62814]: DEBUG oslo_vmware.api [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294926, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.988685] env[62814]: DEBUG oslo_vmware.api [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Task: {'id': task-4294926, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147633} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1367.988940] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1367.989154] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1367.989332] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1367.989504] env[62814]: INFO nova.compute.manager [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1367.989740] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1367.989919] env[62814]: DEBUG nova.compute.manager [-] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1367.990031] env[62814]: DEBUG nova.network.neutron [-] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1368.296157] env[62814]: DEBUG nova.compute.manager [req-94cf1f67-3dd6-48c7-ac09-5b9a421a7732 req-d35ad0c5-c91a-4fc0-b065-074e69c21098 service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Received event network-vif-deleted-cedb024a-64b9-483f-8aa5-1c160162053f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1368.296355] env[62814]: INFO nova.compute.manager [req-94cf1f67-3dd6-48c7-ac09-5b9a421a7732 req-d35ad0c5-c91a-4fc0-b065-074e69c21098 service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Neutron deleted interface cedb024a-64b9-483f-8aa5-1c160162053f; detaching it from the instance and deleting it from the info cache [ 1368.296525] env[62814]: DEBUG nova.network.neutron [req-94cf1f67-3dd6-48c7-ac09-5b9a421a7732 req-d35ad0c5-c91a-4fc0-b065-074e69c21098 service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1368.776356] env[62814]: DEBUG nova.network.neutron [-] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1368.799049] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f73e983-f9d0-48eb-bab0-a91ff44c3b0f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.808932] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67774ce-5d1c-4224-9f0e-322aebc3ed03 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1368.835249] env[62814]: DEBUG nova.compute.manager [req-94cf1f67-3dd6-48c7-ac09-5b9a421a7732 req-d35ad0c5-c91a-4fc0-b065-074e69c21098 service nova] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Detach interface failed, port_id=cedb024a-64b9-483f-8aa5-1c160162053f, reason: Instance 54eb2a99-c7ef-4faf-a0d8-bef576081904 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1369.279072] env[62814]: INFO nova.compute.manager [-] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Took 1.29 seconds to deallocate network for instance. [ 1369.784953] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1369.785364] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1369.785476] env[62814]: DEBUG nova.objects.instance [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lazy-loading 'resources' on Instance uuid 54eb2a99-c7ef-4faf-a0d8-bef576081904 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1370.340220] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f11b8e-5b25-4155-9849-ce77c9dd6c26 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.347465] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d82da44-e652-4e2f-aa1f-c86752a4a91c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.376832] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0fe5a7b-5637-4357-bbda-bf0c25ee60c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.383297] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e105311a-cabf-4cc3-be22-40cc37af9ca2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.395882] env[62814]: DEBUG nova.compute.provider_tree [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1370.899097] env[62814]: DEBUG nova.scheduler.client.report [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1371.404448] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1371.424249] env[62814]: INFO nova.scheduler.client.report [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Deleted allocations for instance 54eb2a99-c7ef-4faf-a0d8-bef576081904 [ 1371.933334] env[62814]: DEBUG oslo_concurrency.lockutils [None req-7879d3f2-270b-49ed-bb01-3b328ee85645 tempest-ServerTagsTestJSON-1397708822 tempest-ServerTagsTestJSON-1397708822-project-member] Lock "54eb2a99-c7ef-4faf-a0d8-bef576081904" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.648s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1384.650993] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1384.651382] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1384.651488] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1384.651668] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1384.651838] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1384.655334] env[62814]: INFO nova.compute.manager [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Terminating instance [ 1385.158744] env[62814]: DEBUG nova.compute.manager [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1385.158976] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1385.159905] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b5d8c7-0399-4b54-bca8-7b0e50faaa9b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.168181] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1385.168405] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-600cf64a-8e8f-4a68-9064-03728e547f21 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1385.174046] env[62814]: DEBUG oslo_vmware.api [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1385.174046] env[62814]: value = "task-4294927" [ 1385.174046] env[62814]: _type = "Task" [ 1385.174046] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1385.181650] env[62814]: DEBUG oslo_vmware.api [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1385.685868] env[62814]: DEBUG oslo_vmware.api [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294927, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.184108] env[62814]: DEBUG oslo_vmware.api [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294927, 'name': PowerOffVM_Task, 'duration_secs': 0.596558} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1386.184401] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1386.184573] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1386.184820] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eeba1d49-5f8a-4e90-9495-cdf55d78e069 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.244209] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1386.244430] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1386.244637] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleting the datastore file [datastore2] ba933282-9711-4e99-ad2d-6ad4c6c516cc {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1386.244932] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-575564a0-94f5-4b39-b8ab-db5e66b20b63 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1386.251103] env[62814]: DEBUG oslo_vmware.api [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for the task: (returnval){ [ 1386.251103] env[62814]: value = "task-4294929" [ 1386.251103] env[62814]: _type = "Task" [ 1386.251103] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1386.258496] env[62814]: DEBUG oslo_vmware.api [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294929, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1386.760701] env[62814]: DEBUG oslo_vmware.api [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Task: {'id': task-4294929, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11747} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1386.761015] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1386.761168] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1386.761352] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1386.761548] env[62814]: INFO nova.compute.manager [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Took 1.60 seconds to destroy the instance on the hypervisor. [ 1386.761807] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1386.761993] env[62814]: DEBUG nova.compute.manager [-] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1386.762099] env[62814]: DEBUG nova.network.neutron [-] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1387.186460] env[62814]: DEBUG nova.compute.manager [req-1955c75c-b42d-4855-8cc7-d71ebf39cee8 req-407d15ef-f480-4fc3-a423-f0e4f276f29f service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Received event network-vif-deleted-b5888ff4-67ff-4165-a2bc-390d621e4efa {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1387.186460] env[62814]: INFO nova.compute.manager [req-1955c75c-b42d-4855-8cc7-d71ebf39cee8 req-407d15ef-f480-4fc3-a423-f0e4f276f29f service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Neutron deleted interface b5888ff4-67ff-4165-a2bc-390d621e4efa; detaching it from the instance and deleting it from the info cache [ 1387.186460] env[62814]: DEBUG nova.network.neutron [req-1955c75c-b42d-4855-8cc7-d71ebf39cee8 req-407d15ef-f480-4fc3-a423-f0e4f276f29f service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1387.423177] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1387.423177] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1387.663859] env[62814]: DEBUG nova.network.neutron [-] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1387.688834] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91e9a62b-f59d-4638-bc71-05fcb9f2f8ce {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.698821] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685653ad-e075-4e18-a523-ec39b393a923 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1387.723127] env[62814]: DEBUG nova.compute.manager [req-1955c75c-b42d-4855-8cc7-d71ebf39cee8 req-407d15ef-f480-4fc3-a423-f0e4f276f29f service nova] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Detach interface failed, port_id=b5888ff4-67ff-4165-a2bc-390d621e4efa, reason: Instance ba933282-9711-4e99-ad2d-6ad4c6c516cc could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1387.925420] env[62814]: DEBUG nova.compute.manager [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1388.169100] env[62814]: INFO nova.compute.manager [-] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Took 1.41 seconds to deallocate network for instance. [ 1388.450780] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1388.451080] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1388.452561] env[62814]: INFO nova.compute.claims [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1388.675474] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1389.506374] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ce191a5-0536-4388-bf21-13b8d87b628f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.514256] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc7bf25-5f8d-4991-aee3-fe041d7324f2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.544642] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9526222d-43bf-4f51-ba34-04bb73e65c13 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.551181] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-473c2b1b-abe2-4bcc-8e7e-184e93c9bd77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1389.563727] env[62814]: DEBUG nova.compute.provider_tree [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1390.067205] env[62814]: DEBUG nova.scheduler.client.report [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1390.572690] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.121s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1390.573243] env[62814]: DEBUG nova.compute.manager [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1390.575799] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.900s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1390.576023] env[62814]: DEBUG nova.objects.instance [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lazy-loading 'resources' on Instance uuid ba933282-9711-4e99-ad2d-6ad4c6c516cc {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1391.081817] env[62814]: DEBUG nova.compute.utils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1391.083334] env[62814]: DEBUG nova.compute.manager [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1391.083506] env[62814]: DEBUG nova.network.neutron [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1391.128742] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc6d5d3e-4faa-4b32-a76b-c2a6b4f78fef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.132393] env[62814]: DEBUG nova.policy [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9516bbf4242140059eeb2182d3e7a58b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb7d25c2161f40db8197b1a56517d7f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1391.138436] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4956d4f-9fbb-4150-8e32-8c91e5deeff0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.170153] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e90e377-fae4-4489-9f3e-ad8e5018e2fb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.177599] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0423dcd5-7a66-421e-8aaf-e53613ad3e89 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1391.190958] env[62814]: DEBUG nova.compute.provider_tree [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1391.398945] env[62814]: DEBUG nova.network.neutron [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Successfully created port: 9490794b-7103-426e-89bd-4a4f328b0a0f {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1391.586263] env[62814]: DEBUG nova.compute.manager [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1391.694518] env[62814]: DEBUG nova.scheduler.client.report [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1392.199381] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.623s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1392.221091] env[62814]: INFO nova.scheduler.client.report [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Deleted allocations for instance ba933282-9711-4e99-ad2d-6ad4c6c516cc [ 1392.596547] env[62814]: DEBUG nova.compute.manager [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1392.624022] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1392.624295] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1392.624452] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1392.624634] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1392.624871] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1392.624934] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1392.625141] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1392.625301] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1392.625468] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1392.625630] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1392.625804] env[62814]: DEBUG nova.virt.hardware [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1392.626764] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f686d6bb-2d3a-480a-96da-c9163acbd86e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.635230] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1c8b2d-fba9-407f-aeac-de0e64d9c777 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1392.730145] env[62814]: DEBUG oslo_concurrency.lockutils [None req-ed4bd749-0fc8-400c-a914-e10797d88664 tempest-AttachVolumeShelveTestJSON-1636104906 tempest-AttachVolumeShelveTestJSON-1636104906-project-member] Lock "ba933282-9711-4e99-ad2d-6ad4c6c516cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.079s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1392.772709] env[62814]: DEBUG nova.compute.manager [req-e87f69fe-5eb4-44e6-aa9d-71389fa9241d req-7d16e89e-8481-4603-99af-1fab5cb8d541 service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Received event network-vif-plugged-9490794b-7103-426e-89bd-4a4f328b0a0f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1392.772989] env[62814]: DEBUG oslo_concurrency.lockutils [req-e87f69fe-5eb4-44e6-aa9d-71389fa9241d req-7d16e89e-8481-4603-99af-1fab5cb8d541 service nova] Acquiring lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1392.773218] env[62814]: DEBUG oslo_concurrency.lockutils [req-e87f69fe-5eb4-44e6-aa9d-71389fa9241d req-7d16e89e-8481-4603-99af-1fab5cb8d541 service nova] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1392.773383] env[62814]: DEBUG oslo_concurrency.lockutils [req-e87f69fe-5eb4-44e6-aa9d-71389fa9241d req-7d16e89e-8481-4603-99af-1fab5cb8d541 service nova] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1392.773628] env[62814]: DEBUG nova.compute.manager [req-e87f69fe-5eb4-44e6-aa9d-71389fa9241d req-7d16e89e-8481-4603-99af-1fab5cb8d541 service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] No waiting events found dispatching network-vif-plugged-9490794b-7103-426e-89bd-4a4f328b0a0f {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1392.773741] env[62814]: WARNING nova.compute.manager [req-e87f69fe-5eb4-44e6-aa9d-71389fa9241d req-7d16e89e-8481-4603-99af-1fab5cb8d541 service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Received unexpected event network-vif-plugged-9490794b-7103-426e-89bd-4a4f328b0a0f for instance with vm_state building and task_state spawning. [ 1392.854925] env[62814]: DEBUG nova.network.neutron [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Successfully updated port: 9490794b-7103-426e-89bd-4a4f328b0a0f {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1393.358552] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "refresh_cache-1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1393.358669] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "refresh_cache-1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1393.358826] env[62814]: DEBUG nova.network.neutron [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1393.897604] env[62814]: DEBUG nova.network.neutron [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1394.022266] env[62814]: DEBUG nova.network.neutron [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Updating instance_info_cache with network_info: [{"id": "9490794b-7103-426e-89bd-4a4f328b0a0f", "address": "fa:16:3e:b3:6d:43", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9490794b-71", "ovs_interfaceid": "9490794b-7103-426e-89bd-4a4f328b0a0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1394.525214] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "refresh_cache-1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1394.525559] env[62814]: DEBUG nova.compute.manager [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Instance network_info: |[{"id": "9490794b-7103-426e-89bd-4a4f328b0a0f", "address": "fa:16:3e:b3:6d:43", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9490794b-71", "ovs_interfaceid": "9490794b-7103-426e-89bd-4a4f328b0a0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1394.526021] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:6d:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ffb921-796a-40fe-9662-d3fc01547dcb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9490794b-7103-426e-89bd-4a4f328b0a0f', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1394.533453] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1394.533692] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1394.533947] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b7d778d-029c-42e5-a7e1-afaefd5f44b5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1394.552218] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1394.552218] env[62814]: value = "task-4294931" [ 1394.552218] env[62814]: _type = "Task" [ 1394.552218] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1394.559223] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294931, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1394.800205] env[62814]: DEBUG nova.compute.manager [req-1ec45450-64c8-4e4e-ade5-8e1222d31051 req-242388a1-19d4-4995-b460-198c58f00ce9 service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Received event network-changed-9490794b-7103-426e-89bd-4a4f328b0a0f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1394.800397] env[62814]: DEBUG nova.compute.manager [req-1ec45450-64c8-4e4e-ade5-8e1222d31051 req-242388a1-19d4-4995-b460-198c58f00ce9 service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Refreshing instance network info cache due to event network-changed-9490794b-7103-426e-89bd-4a4f328b0a0f. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1394.800774] env[62814]: DEBUG oslo_concurrency.lockutils [req-1ec45450-64c8-4e4e-ade5-8e1222d31051 req-242388a1-19d4-4995-b460-198c58f00ce9 service nova] Acquiring lock "refresh_cache-1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1394.800995] env[62814]: DEBUG oslo_concurrency.lockutils [req-1ec45450-64c8-4e4e-ade5-8e1222d31051 req-242388a1-19d4-4995-b460-198c58f00ce9 service nova] Acquired lock "refresh_cache-1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1394.801203] env[62814]: DEBUG nova.network.neutron [req-1ec45450-64c8-4e4e-ade5-8e1222d31051 req-242388a1-19d4-4995-b460-198c58f00ce9 service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Refreshing network info cache for port 9490794b-7103-426e-89bd-4a4f328b0a0f {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1395.062661] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294931, 'name': CreateVM_Task, 'duration_secs': 0.268354} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1395.062935] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1395.063528] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1395.063700] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1395.064027] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1395.064295] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef8aa188-802d-4047-9540-157e2d89498b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.068800] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1395.068800] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d13978-1203-dc85-e3c9-b24dbaac12d0" [ 1395.068800] env[62814]: _type = "Task" [ 1395.068800] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1395.076185] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d13978-1203-dc85-e3c9-b24dbaac12d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1395.508752] env[62814]: DEBUG nova.network.neutron [req-1ec45450-64c8-4e4e-ade5-8e1222d31051 req-242388a1-19d4-4995-b460-198c58f00ce9 service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Updated VIF entry in instance network info cache for port 9490794b-7103-426e-89bd-4a4f328b0a0f. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1395.509171] env[62814]: DEBUG nova.network.neutron [req-1ec45450-64c8-4e4e-ade5-8e1222d31051 req-242388a1-19d4-4995-b460-198c58f00ce9 service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Updating instance_info_cache with network_info: [{"id": "9490794b-7103-426e-89bd-4a4f328b0a0f", "address": "fa:16:3e:b3:6d:43", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9490794b-71", "ovs_interfaceid": "9490794b-7103-426e-89bd-4a4f328b0a0f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1395.578288] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d13978-1203-dc85-e3c9-b24dbaac12d0, 'name': SearchDatastore_Task, 'duration_secs': 0.010021} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1395.578585] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1395.578818] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1395.579058] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1395.579209] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1395.579384] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1395.579635] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24825c17-a18e-4181-8bcf-115783929818 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.587577] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1395.587710] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1395.588385] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9865caae-7507-450f-97b7-ddd8c3043988 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1395.592868] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1395.592868] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d0d201-f09f-fe9d-3305-9be6bffba2e5" [ 1395.592868] env[62814]: _type = "Task" [ 1395.592868] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1395.599568] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d0d201-f09f-fe9d-3305-9be6bffba2e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1396.012248] env[62814]: DEBUG oslo_concurrency.lockutils [req-1ec45450-64c8-4e4e-ade5-8e1222d31051 req-242388a1-19d4-4995-b460-198c58f00ce9 service nova] Releasing lock "refresh_cache-1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1396.103625] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d0d201-f09f-fe9d-3305-9be6bffba2e5, 'name': SearchDatastore_Task, 'duration_secs': 0.009883} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1396.104402] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4192ba6d-11d9-448b-ac97-87e28ed2b1dd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.109629] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1396.109629] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e968ee-f4a7-c09d-966b-1239fbe3335e" [ 1396.109629] env[62814]: _type = "Task" [ 1396.109629] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.116902] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e968ee-f4a7-c09d-966b-1239fbe3335e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1396.621460] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52e968ee-f4a7-c09d-966b-1239fbe3335e, 'name': SearchDatastore_Task, 'duration_secs': 0.010532} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1396.621875] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1396.622167] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9/1a7fb769-f0bf-4387-a3a8-69b05d18f5c9.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1396.623048] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fdc5140-4b9d-46ad-acd7-9ac6b9204d7d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1396.629391] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1396.629391] env[62814]: value = "task-4294932" [ 1396.629391] env[62814]: _type = "Task" [ 1396.629391] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1396.638775] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294932, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.139194] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294932, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446156} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.139521] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9/1a7fb769-f0bf-4387-a3a8-69b05d18f5c9.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1397.139669] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1397.139874] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25d16737-8109-435a-930c-03b56cfff829 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.145756] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1397.145756] env[62814]: value = "task-4294933" [ 1397.145756] env[62814]: _type = "Task" [ 1397.145756] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1397.152488] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294933, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1397.655712] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294933, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058024} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1397.655923] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1397.656695] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5f0cef-5f41-486f-8a1c-79de704f1434 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.677804] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Reconfiguring VM instance instance-0000007c to attach disk [datastore2] 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9/1a7fb769-f0bf-4387-a3a8-69b05d18f5c9.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1397.678475] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5da7866d-0ec6-46d2-a76b-8c8d5c353940 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1397.697348] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1397.697348] env[62814]: value = "task-4294934" [ 1397.697348] env[62814]: _type = "Task" [ 1397.697348] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1397.705112] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294934, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1398.208256] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294934, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1398.709163] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294934, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1399.210044] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294934, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1399.710251] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294934, 'name': ReconfigVM_Task, 'duration_secs': 1.779429} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1399.710545] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Reconfigured VM instance instance-0000007c to attach disk [datastore2] 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9/1a7fb769-f0bf-4387-a3a8-69b05d18f5c9.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1399.711270] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2802fe30-b881-45ca-9ae3-cd9fed1a7717 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1399.717073] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1399.717073] env[62814]: value = "task-4294935" [ 1399.717073] env[62814]: _type = "Task" [ 1399.717073] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1399.724197] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294935, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1400.227565] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294935, 'name': Rename_Task, 'duration_secs': 0.147935} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1400.227928] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1400.227928] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0c1aa11a-28a9-429e-a8cb-a918e30a40fe {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1400.233960] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1400.233960] env[62814]: value = "task-4294936" [ 1400.233960] env[62814]: _type = "Task" [ 1400.233960] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1400.241061] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294936, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1400.746645] env[62814]: DEBUG oslo_vmware.api [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294936, 'name': PowerOnVM_Task, 'duration_secs': 0.413047} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1400.746992] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1400.747285] env[62814]: INFO nova.compute.manager [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Took 8.15 seconds to spawn the instance on the hypervisor. [ 1400.747534] env[62814]: DEBUG nova.compute.manager [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1400.748718] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-928a2102-d632-475d-8d13-b7171e33dfb2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.268107] env[62814]: INFO nova.compute.manager [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Took 12.84 seconds to build instance. [ 1401.770277] env[62814]: DEBUG oslo_concurrency.lockutils [None req-c159daa0-0043-42f4-9354-24faeb64516b tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.348s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1401.863072] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1401.863279] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1401.863444] env[62814]: DEBUG nova.compute.manager [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1401.864449] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db00efdb-a304-4396-8d97-1bf622a3213b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1401.871455] env[62814]: DEBUG nova.compute.manager [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1401.872030] env[62814]: DEBUG nova.objects.instance [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lazy-loading 'flavor' on Instance uuid 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1402.879623] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1402.879971] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a7e7f48c-f23b-468a-abd1-788fa2e1ed5c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1402.886154] env[62814]: DEBUG oslo_vmware.api [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1402.886154] env[62814]: value = "task-4294937" [ 1402.886154] env[62814]: _type = "Task" [ 1402.886154] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1402.893470] env[62814]: DEBUG oslo_vmware.api [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1403.395448] env[62814]: DEBUG oslo_vmware.api [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294937, 'name': PowerOffVM_Task, 'duration_secs': 0.172478} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1403.395722] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1403.395920] env[62814]: DEBUG nova.compute.manager [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1403.396708] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5556a8-cc67-4806-9624-a470724f719c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1403.907241] env[62814]: DEBUG oslo_concurrency.lockutils [None req-0e1462d3-bc49-4215-bbad-9df6b62d13cb tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.044s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1404.333662] env[62814]: INFO nova.compute.manager [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Rebuilding instance [ 1404.374340] env[62814]: DEBUG nova.compute.manager [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1404.375225] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6573fb62-359e-47c7-acd0-6bffd2f6084b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.389446] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1405.389806] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5141f728-84ce-4c55-8dbe-1ecc6bad3cb1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.396627] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1405.396627] env[62814]: value = "task-4294938" [ 1405.396627] env[62814]: _type = "Task" [ 1405.396627] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.403869] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294938, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.906668] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1405.906911] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1405.907658] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f71f898-9eba-40fd-8494-f7c50ca96314 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.914008] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1405.914224] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-072da7f7-b2c8-4fd1-b72f-f667eaaffde3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.978080] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1405.978314] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1405.978508] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleting the datastore file [datastore2] 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1405.978798] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-05cf575e-9409-4d1f-968a-b317adb2e579 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.985339] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1405.985339] env[62814]: value = "task-4294940" [ 1405.985339] env[62814]: _type = "Task" [ 1405.985339] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1405.992869] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294940, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1406.495209] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294940, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138724} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1406.495479] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1406.495639] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1406.495809] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1407.502866] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.506067] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.506067] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 1407.506481] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1407.528400] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1407.528537] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1407.528645] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1407.529333] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1407.529333] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1407.529333] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1407.529460] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1407.529570] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1407.529684] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1407.529984] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1407.531019] env[62814]: DEBUG nova.virt.hardware [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1407.531331] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1df29b5-3654-45ed-a4a5-a774bcff9c34 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.539113] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d193ac81-590f-4b84-bd62-a37d20b943e8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.552618] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:6d:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ffb921-796a-40fe-9662-d3fc01547dcb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9490794b-7103-426e-89bd-4a4f328b0a0f', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1407.559759] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1407.560308] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1407.560623] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c184850-67c9-4362-ab6a-1186e5ee3bd9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1407.579831] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1407.579831] env[62814]: value = "task-4294941" [ 1407.579831] env[62814]: _type = "Task" [ 1407.579831] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1407.586846] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294941, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.010046] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1408.010046] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1408.010046] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1408.010282] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1408.011077] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6329d791-5f21-4f86-bc54-9e35e2e4b6f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.019047] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca254479-230d-4cb5-9911-37c2f8703a8b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.033481] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c4cb76-5a97-4488-a4c0-ef88319fb086 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.039789] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1193ce-e76d-4022-8aae-3c134948f1cf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.067736] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180479MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1408.068054] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1408.068374] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1408.087920] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294941, 'name': CreateVM_Task, 'duration_secs': 0.296876} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1408.088331] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1408.091855] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1408.091855] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1408.091855] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1408.091855] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a554654-98d4-4c4a-aef4-ee59727ec7c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.095110] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1408.095110] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5226cb06-f491-3fd0-ae2b-4432f7bd618c" [ 1408.095110] env[62814]: _type = "Task" [ 1408.095110] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1408.102658] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5226cb06-f491-3fd0-ae2b-4432f7bd618c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1408.606048] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5226cb06-f491-3fd0-ae2b-4432f7bd618c, 'name': SearchDatastore_Task, 'duration_secs': 0.01002} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1408.606048] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1408.606048] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1408.606048] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1408.606048] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1408.606568] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1408.606568] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2de4b287-58ef-4efc-be87-67c8fd05a973 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.614974] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1408.615343] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1408.616396] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27001cf2-da13-4713-87ac-9966af1812fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1408.621253] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1408.621253] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528c974c-6d40-aa3a-153d-5b07b4b7bb35" [ 1408.621253] env[62814]: _type = "Task" [ 1408.621253] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1408.630280] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528c974c-6d40-aa3a-153d-5b07b4b7bb35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.095258] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 711f9623-4c61-47da-b7e5-15a41487b3b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1409.095420] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1409.095631] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1409.095780] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=149GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1409.130700] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]528c974c-6d40-aa3a-153d-5b07b4b7bb35, 'name': SearchDatastore_Task, 'duration_secs': 0.008528} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1409.132501] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575e964c-be29-4c74-8e11-fdff9ba11ea3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.134739] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e82f440-3205-4a0f-b99c-1886002593ac {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.142819] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da55a8dc-d6f2-4b0a-873f-2d64bf984a73 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.146257] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1409.146257] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ad3a6c-d8d4-6c69-fd47-f5aa6893ed31" [ 1409.146257] env[62814]: _type = "Task" [ 1409.146257] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1409.172937] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88eeaa67-8cff-4f9a-ae29-917109b16091 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.178215] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ad3a6c-d8d4-6c69-fd47-f5aa6893ed31, 'name': SearchDatastore_Task, 'duration_secs': 0.00955} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1409.178782] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1409.179118] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9/1a7fb769-f0bf-4387-a3a8-69b05d18f5c9.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1409.179330] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2f52fc6a-15de-4bed-881f-68118241f117 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.184075] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20265c0-c7bb-4661-b965-51c5389c4662 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.188351] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1409.188351] env[62814]: value = "task-4294942" [ 1409.188351] env[62814]: _type = "Task" [ 1409.188351] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1409.199283] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1409.204629] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294942, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.698473] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294942, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447654} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1409.698855] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9/1a7fb769-f0bf-4387-a3a8-69b05d18f5c9.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1409.698931] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1409.699164] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9abca2fa-f2de-4413-a2a8-cfcde5ebe7b3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.702753] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1409.707545] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1409.707545] env[62814]: value = "task-4294943" [ 1409.707545] env[62814]: _type = "Task" [ 1409.707545] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1409.715449] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294943, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.207920] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1410.208135] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.140s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1410.216325] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294943, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.056751} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1410.216589] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1410.217341] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb01b7e-9596-41b2-80bc-92a180e1e4f5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.240015] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Reconfiguring VM instance instance-0000007c to attach disk [datastore2] 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9/1a7fb769-f0bf-4387-a3a8-69b05d18f5c9.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1410.240544] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b6d1d38-83f7-4ded-b491-e6dfa42f1b6e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.259141] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1410.259141] env[62814]: value = "task-4294944" [ 1410.259141] env[62814]: _type = "Task" [ 1410.259141] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1410.266480] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294944, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1410.768559] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294944, 'name': ReconfigVM_Task, 'duration_secs': 0.271516} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1410.768950] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Reconfigured VM instance instance-0000007c to attach disk [datastore2] 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9/1a7fb769-f0bf-4387-a3a8-69b05d18f5c9.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1410.769457] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-646e78a1-e1d6-48be-9c10-9d265e55a78e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1410.776148] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1410.776148] env[62814]: value = "task-4294945" [ 1410.776148] env[62814]: _type = "Task" [ 1410.776148] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1410.783146] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294945, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.286380] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294945, 'name': Rename_Task, 'duration_secs': 0.131387} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1411.286704] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1411.286943] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09afc705-f878-48f4-b8ba-1d765174ec85 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.293022] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1411.293022] env[62814]: value = "task-4294946" [ 1411.293022] env[62814]: _type = "Task" [ 1411.293022] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.299909] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294946, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.802755] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294946, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1412.210054] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1412.210054] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1412.210054] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1412.210054] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1412.210054] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1412.303163] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294946, 'name': PowerOnVM_Task, 'duration_secs': 0.722895} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1412.303432] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1412.303640] env[62814]: DEBUG nova.compute.manager [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1412.304409] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15f71632-fdf2-463e-869b-53bf1dcba185 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.815732] env[62814]: INFO nova.compute.manager [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] bringing vm to original state: 'stopped' [ 1413.502436] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1413.822885] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1413.823167] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1413.823353] env[62814]: DEBUG nova.compute.manager [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1413.824280] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b23cad6-dc3d-4ed9-8720-d808a55a1bd8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.831459] env[62814]: DEBUG nova.compute.manager [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62814) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3404}} [ 1414.337979] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1414.338339] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca87b25f-30c0-43fd-ab40-03577b842908 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1414.346376] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1414.346376] env[62814]: value = "task-4294947" [ 1414.346376] env[62814]: _type = "Task" [ 1414.346376] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1414.354051] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294947, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1414.856440] env[62814]: DEBUG oslo_vmware.api [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294947, 'name': PowerOffVM_Task, 'duration_secs': 0.167895} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1414.856819] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1414.856896] env[62814]: DEBUG nova.compute.manager [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1414.857605] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dedbde7a-0952-48b0-bbb9-43a87cfd81cb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.368932] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.546s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1415.787321] env[62814]: DEBUG oslo_concurrency.lockutils [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1415.787596] env[62814]: DEBUG oslo_concurrency.lockutils [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1415.787806] env[62814]: DEBUG oslo_concurrency.lockutils [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1415.787988] env[62814]: DEBUG oslo_concurrency.lockutils [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1415.788173] env[62814]: DEBUG oslo_concurrency.lockutils [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1415.790251] env[62814]: INFO nova.compute.manager [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Terminating instance [ 1415.875607] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1415.875977] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1415.876036] env[62814]: DEBUG nova.objects.instance [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1416.293533] env[62814]: DEBUG nova.compute.manager [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1416.293833] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1416.294735] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8875027e-fd28-41bc-ad16-3c3d5e816cec {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.302468] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1416.302687] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c2ccbabc-743d-4c5d-bde6-ba1a9c04b29a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.365927] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1416.366147] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1416.366325] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleting the datastore file [datastore2] 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1416.366614] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f5e4da0-9864-43e4-b45c-fc916ee9c6c6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1416.373244] env[62814]: DEBUG oslo_vmware.api [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1416.373244] env[62814]: value = "task-4294949" [ 1416.373244] env[62814]: _type = "Task" [ 1416.373244] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1416.383648] env[62814]: DEBUG oslo_vmware.api [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294949, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.882908] env[62814]: DEBUG oslo_vmware.api [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294949, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125985} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.883295] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1416.883336] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1416.883518] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1416.883690] env[62814]: INFO nova.compute.manager [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1416.883925] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1416.884119] env[62814]: DEBUG nova.compute.manager [-] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1416.884217] env[62814]: DEBUG nova.network.neutron [-] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1416.886523] env[62814]: DEBUG oslo_concurrency.lockutils [None req-bf1dc885-6244-4a4f-8982-e7c5b91569e0 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1417.176218] env[62814]: DEBUG nova.compute.manager [req-27b58d29-0f3d-437f-86b1-fe0dac16678c req-07727ed8-3265-4b9a-9a56-5464ebbef3ee service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Received event network-vif-deleted-9490794b-7103-426e-89bd-4a4f328b0a0f {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1417.176378] env[62814]: INFO nova.compute.manager [req-27b58d29-0f3d-437f-86b1-fe0dac16678c req-07727ed8-3265-4b9a-9a56-5464ebbef3ee service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Neutron deleted interface 9490794b-7103-426e-89bd-4a4f328b0a0f; detaching it from the instance and deleting it from the info cache [ 1417.176582] env[62814]: DEBUG nova.network.neutron [req-27b58d29-0f3d-437f-86b1-fe0dac16678c req-07727ed8-3265-4b9a-9a56-5464ebbef3ee service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1417.653824] env[62814]: DEBUG nova.network.neutron [-] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1417.678442] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dad9b7b5-4d51-4eb7-9375-17e4ff38f767 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.688948] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5944a448-d5d2-4509-a76b-0e533a67cdf0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.712833] env[62814]: DEBUG nova.compute.manager [req-27b58d29-0f3d-437f-86b1-fe0dac16678c req-07727ed8-3265-4b9a-9a56-5464ebbef3ee service nova] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Detach interface failed, port_id=9490794b-7103-426e-89bd-4a4f328b0a0f, reason: Instance 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1418.157565] env[62814]: INFO nova.compute.manager [-] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Took 1.27 seconds to deallocate network for instance. [ 1418.664265] env[62814]: DEBUG oslo_concurrency.lockutils [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1418.664640] env[62814]: DEBUG oslo_concurrency.lockutils [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1418.664882] env[62814]: DEBUG nova.objects.instance [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lazy-loading 'resources' on Instance uuid 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1419.209431] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8a98e9d-3876-42fe-869c-41ffb8303093 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.216950] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e313fc74-1278-4706-974f-16a47cc44bcb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.246608] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb39622b-5262-4b17-b1b8-c430e5fc0e36 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.253299] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a081e1-e5d0-4a8e-a9b8-58337df781e2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.267145] env[62814]: DEBUG nova.compute.provider_tree [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1419.770872] env[62814]: DEBUG nova.scheduler.client.report [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1420.276633] env[62814]: DEBUG oslo_concurrency.lockutils [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.612s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1420.300954] env[62814]: INFO nova.scheduler.client.report [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted allocations for instance 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9 [ 1420.809030] env[62814]: DEBUG oslo_concurrency.lockutils [None req-83527d6e-5b55-4034-9320-033717afb341 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "1a7fb769-f0bf-4387-a3a8-69b05d18f5c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.021s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1422.759028] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1422.759380] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1423.262189] env[62814]: DEBUG nova.compute.manager [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1423.785598] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1423.785902] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1423.787415] env[62814]: INFO nova.compute.claims [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1424.830988] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7969891-49ab-40d8-893b-c54608e96e65 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.838435] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecedbe52-b110-44a5-bedc-18c997a4b7df {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.868223] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8145f09-ec8d-4156-a8b8-6787ef3db042 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.875068] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8308b2b9-566c-49e0-a005-d9474aab9c1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1424.887480] env[62814]: DEBUG nova.compute.provider_tree [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1425.390274] env[62814]: DEBUG nova.scheduler.client.report [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1425.894646] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.109s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1425.895224] env[62814]: DEBUG nova.compute.manager [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1426.400717] env[62814]: DEBUG nova.compute.utils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1426.402089] env[62814]: DEBUG nova.compute.manager [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1426.402261] env[62814]: DEBUG nova.network.neutron [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1426.446853] env[62814]: DEBUG nova.policy [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9516bbf4242140059eeb2182d3e7a58b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb7d25c2161f40db8197b1a56517d7f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1426.771740] env[62814]: DEBUG nova.network.neutron [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Successfully created port: b03c4787-0abf-456e-8f11-e9ae79f8c6ab {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1426.905171] env[62814]: DEBUG nova.compute.manager [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1427.916130] env[62814]: DEBUG nova.compute.manager [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1427.941375] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1427.941628] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1427.941785] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1427.941971] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1427.942132] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1427.942281] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1427.942579] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1427.942754] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1427.942922] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1427.943105] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1427.943280] env[62814]: DEBUG nova.virt.hardware [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1427.944139] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777da04f-02a5-4efe-ac43-1928db7ec92f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1427.951626] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c40f4d31-9b23-48fc-b4f1-5f630fc51667 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1428.123725] env[62814]: DEBUG nova.compute.manager [req-29acdf1e-9d2f-48ae-97ef-3ab907e37be9 req-80d6bd5c-e419-4dc8-ac43-221ab045006e service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Received event network-vif-plugged-b03c4787-0abf-456e-8f11-e9ae79f8c6ab {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1428.123950] env[62814]: DEBUG oslo_concurrency.lockutils [req-29acdf1e-9d2f-48ae-97ef-3ab907e37be9 req-80d6bd5c-e419-4dc8-ac43-221ab045006e service nova] Acquiring lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1428.124189] env[62814]: DEBUG oslo_concurrency.lockutils [req-29acdf1e-9d2f-48ae-97ef-3ab907e37be9 req-80d6bd5c-e419-4dc8-ac43-221ab045006e service nova] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1428.124329] env[62814]: DEBUG oslo_concurrency.lockutils [req-29acdf1e-9d2f-48ae-97ef-3ab907e37be9 req-80d6bd5c-e419-4dc8-ac43-221ab045006e service nova] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1428.124493] env[62814]: DEBUG nova.compute.manager [req-29acdf1e-9d2f-48ae-97ef-3ab907e37be9 req-80d6bd5c-e419-4dc8-ac43-221ab045006e service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] No waiting events found dispatching network-vif-plugged-b03c4787-0abf-456e-8f11-e9ae79f8c6ab {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1428.124673] env[62814]: WARNING nova.compute.manager [req-29acdf1e-9d2f-48ae-97ef-3ab907e37be9 req-80d6bd5c-e419-4dc8-ac43-221ab045006e service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Received unexpected event network-vif-plugged-b03c4787-0abf-456e-8f11-e9ae79f8c6ab for instance with vm_state building and task_state spawning. [ 1428.207060] env[62814]: DEBUG nova.network.neutron [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Successfully updated port: b03c4787-0abf-456e-8f11-e9ae79f8c6ab {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1428.710201] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "refresh_cache-63b4c501-f3c7-4366-bd3e-83ba769f6f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1428.710391] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "refresh_cache-63b4c501-f3c7-4366-bd3e-83ba769f6f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1428.710705] env[62814]: DEBUG nova.network.neutron [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1429.242553] env[62814]: DEBUG nova.network.neutron [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1429.374548] env[62814]: DEBUG nova.network.neutron [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Updating instance_info_cache with network_info: [{"id": "b03c4787-0abf-456e-8f11-e9ae79f8c6ab", "address": "fa:16:3e:0c:bc:bc", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb03c4787-0a", "ovs_interfaceid": "b03c4787-0abf-456e-8f11-e9ae79f8c6ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1429.877683] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "refresh_cache-63b4c501-f3c7-4366-bd3e-83ba769f6f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1429.878025] env[62814]: DEBUG nova.compute.manager [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Instance network_info: |[{"id": "b03c4787-0abf-456e-8f11-e9ae79f8c6ab", "address": "fa:16:3e:0c:bc:bc", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb03c4787-0a", "ovs_interfaceid": "b03c4787-0abf-456e-8f11-e9ae79f8c6ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1429.878448] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:bc:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ffb921-796a-40fe-9662-d3fc01547dcb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b03c4787-0abf-456e-8f11-e9ae79f8c6ab', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1429.885869] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1429.886047] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1429.886267] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53c34a68-9426-4d5d-9104-4468961450d4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1429.905845] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1429.905845] env[62814]: value = "task-4294950" [ 1429.905845] env[62814]: _type = "Task" [ 1429.905845] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1429.913168] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294950, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.150663] env[62814]: DEBUG nova.compute.manager [req-444af67f-325b-4028-9c22-998fec2357cd req-19f0ded7-2a76-49ab-b42a-21dd48766ff6 service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Received event network-changed-b03c4787-0abf-456e-8f11-e9ae79f8c6ab {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1430.150874] env[62814]: DEBUG nova.compute.manager [req-444af67f-325b-4028-9c22-998fec2357cd req-19f0ded7-2a76-49ab-b42a-21dd48766ff6 service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Refreshing instance network info cache due to event network-changed-b03c4787-0abf-456e-8f11-e9ae79f8c6ab. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1430.151101] env[62814]: DEBUG oslo_concurrency.lockutils [req-444af67f-325b-4028-9c22-998fec2357cd req-19f0ded7-2a76-49ab-b42a-21dd48766ff6 service nova] Acquiring lock "refresh_cache-63b4c501-f3c7-4366-bd3e-83ba769f6f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1430.151246] env[62814]: DEBUG oslo_concurrency.lockutils [req-444af67f-325b-4028-9c22-998fec2357cd req-19f0ded7-2a76-49ab-b42a-21dd48766ff6 service nova] Acquired lock "refresh_cache-63b4c501-f3c7-4366-bd3e-83ba769f6f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1430.151404] env[62814]: DEBUG nova.network.neutron [req-444af67f-325b-4028-9c22-998fec2357cd req-19f0ded7-2a76-49ab-b42a-21dd48766ff6 service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Refreshing network info cache for port b03c4787-0abf-456e-8f11-e9ae79f8c6ab {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1430.415609] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294950, 'name': CreateVM_Task, 'duration_secs': 0.281948} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.415978] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1430.416477] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1430.416679] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1430.417035] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1430.417283] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d654e7d-8fd5-4ee7-9aed-a57934126078 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.422290] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1430.422290] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527d3f82-79cc-fde0-0306-001f728eefe4" [ 1430.422290] env[62814]: _type = "Task" [ 1430.422290] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1430.429446] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527d3f82-79cc-fde0-0306-001f728eefe4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1430.862392] env[62814]: DEBUG nova.network.neutron [req-444af67f-325b-4028-9c22-998fec2357cd req-19f0ded7-2a76-49ab-b42a-21dd48766ff6 service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Updated VIF entry in instance network info cache for port b03c4787-0abf-456e-8f11-e9ae79f8c6ab. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1430.862762] env[62814]: DEBUG nova.network.neutron [req-444af67f-325b-4028-9c22-998fec2357cd req-19f0ded7-2a76-49ab-b42a-21dd48766ff6 service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Updating instance_info_cache with network_info: [{"id": "b03c4787-0abf-456e-8f11-e9ae79f8c6ab", "address": "fa:16:3e:0c:bc:bc", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb03c4787-0a", "ovs_interfaceid": "b03c4787-0abf-456e-8f11-e9ae79f8c6ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1430.932325] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]527d3f82-79cc-fde0-0306-001f728eefe4, 'name': SearchDatastore_Task, 'duration_secs': 0.0097} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1430.932569] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1430.932796] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1430.933039] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1430.933194] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1430.933377] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1430.933623] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75608f5a-5249-4446-ba3e-cc515d6d9c78 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.941923] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1430.942019] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1430.942661] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a97c605-559f-4197-bc2b-14ae36c2acaf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1430.947352] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1430.947352] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523dee21-f5cf-211a-3c4a-25feb8f2f3a4" [ 1430.947352] env[62814]: _type = "Task" [ 1430.947352] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1430.954010] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523dee21-f5cf-211a-3c4a-25feb8f2f3a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1431.365599] env[62814]: DEBUG oslo_concurrency.lockutils [req-444af67f-325b-4028-9c22-998fec2357cd req-19f0ded7-2a76-49ab-b42a-21dd48766ff6 service nova] Releasing lock "refresh_cache-63b4c501-f3c7-4366-bd3e-83ba769f6f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1431.456990] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]523dee21-f5cf-211a-3c4a-25feb8f2f3a4, 'name': SearchDatastore_Task, 'duration_secs': 0.008141} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1431.457762] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc1157cb-0d62-49e8-a762-01ed09764266 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.463229] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1431.463229] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5222c573-000a-a6e1-cfd1-30934e68f909" [ 1431.463229] env[62814]: _type = "Task" [ 1431.463229] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1431.470101] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5222c573-000a-a6e1-cfd1-30934e68f909, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1431.973521] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5222c573-000a-a6e1-cfd1-30934e68f909, 'name': SearchDatastore_Task, 'duration_secs': 0.009224} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1431.973780] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1431.974062] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 63b4c501-f3c7-4366-bd3e-83ba769f6f9f/63b4c501-f3c7-4366-bd3e-83ba769f6f9f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1431.974313] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ed67ea5-f3d2-4e82-922a-7bdc1f5d5d38 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1431.980329] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1431.980329] env[62814]: value = "task-4294951" [ 1431.980329] env[62814]: _type = "Task" [ 1431.980329] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1431.987390] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294951, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1432.489819] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294951, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453686} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1432.490178] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 63b4c501-f3c7-4366-bd3e-83ba769f6f9f/63b4c501-f3c7-4366-bd3e-83ba769f6f9f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1432.490335] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1432.490581] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fde19d42-53cf-49cb-afe0-8be963de5f8a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1432.496986] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1432.496986] env[62814]: value = "task-4294952" [ 1432.496986] env[62814]: _type = "Task" [ 1432.496986] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1432.503676] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294952, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.006690] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294952, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063819} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1433.008036] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1433.008036] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9306d0-fedb-455f-b30e-989324787c83 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.028907] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] 63b4c501-f3c7-4366-bd3e-83ba769f6f9f/63b4c501-f3c7-4366-bd3e-83ba769f6f9f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1433.029157] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2458da9-a32e-4e3c-9e01-930a2d2b1bb3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.048974] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1433.048974] env[62814]: value = "task-4294953" [ 1433.048974] env[62814]: _type = "Task" [ 1433.048974] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1433.056394] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294953, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1433.558018] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294953, 'name': ReconfigVM_Task, 'duration_secs': 0.321295} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1433.558363] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfigured VM instance instance-0000007d to attach disk [datastore2] 63b4c501-f3c7-4366-bd3e-83ba769f6f9f/63b4c501-f3c7-4366-bd3e-83ba769f6f9f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1433.558942] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e447e1b3-3d8e-41e7-9c9a-1162ec2b7514 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1433.565062] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1433.565062] env[62814]: value = "task-4294954" [ 1433.565062] env[62814]: _type = "Task" [ 1433.565062] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1433.572185] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294954, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.074235] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294954, 'name': Rename_Task, 'duration_secs': 0.139318} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1434.074559] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1434.074855] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94711c1e-232c-4bd9-a3dc-baa545d18116 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1434.081980] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1434.081980] env[62814]: value = "task-4294955" [ 1434.081980] env[62814]: _type = "Task" [ 1434.081980] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1434.090467] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294955, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1434.591406] env[62814]: DEBUG oslo_vmware.api [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294955, 'name': PowerOnVM_Task, 'duration_secs': 0.418024} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1434.591850] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1434.591850] env[62814]: INFO nova.compute.manager [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Took 6.68 seconds to spawn the instance on the hypervisor. [ 1434.592028] env[62814]: DEBUG nova.compute.manager [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1434.592763] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fba1d25-0a61-4816-bb2d-10828bc623e7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.109611] env[62814]: INFO nova.compute.manager [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Took 11.34 seconds to build instance. [ 1435.612017] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d68226ae-2c9b-4fd9-ba10-62b93c1e905f tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.853s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1435.756824] env[62814]: DEBUG nova.compute.manager [req-28a9102a-d572-4462-a40f-c01de284f55d req-ca2caccc-d591-4ed7-860e-fb22737b3b9e service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Received event network-changed-b03c4787-0abf-456e-8f11-e9ae79f8c6ab {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1435.757068] env[62814]: DEBUG nova.compute.manager [req-28a9102a-d572-4462-a40f-c01de284f55d req-ca2caccc-d591-4ed7-860e-fb22737b3b9e service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Refreshing instance network info cache due to event network-changed-b03c4787-0abf-456e-8f11-e9ae79f8c6ab. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1435.757306] env[62814]: DEBUG oslo_concurrency.lockutils [req-28a9102a-d572-4462-a40f-c01de284f55d req-ca2caccc-d591-4ed7-860e-fb22737b3b9e service nova] Acquiring lock "refresh_cache-63b4c501-f3c7-4366-bd3e-83ba769f6f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1435.757519] env[62814]: DEBUG oslo_concurrency.lockutils [req-28a9102a-d572-4462-a40f-c01de284f55d req-ca2caccc-d591-4ed7-860e-fb22737b3b9e service nova] Acquired lock "refresh_cache-63b4c501-f3c7-4366-bd3e-83ba769f6f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1435.757712] env[62814]: DEBUG nova.network.neutron [req-28a9102a-d572-4462-a40f-c01de284f55d req-ca2caccc-d591-4ed7-860e-fb22737b3b9e service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Refreshing network info cache for port b03c4787-0abf-456e-8f11-e9ae79f8c6ab {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1436.453186] env[62814]: DEBUG nova.network.neutron [req-28a9102a-d572-4462-a40f-c01de284f55d req-ca2caccc-d591-4ed7-860e-fb22737b3b9e service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Updated VIF entry in instance network info cache for port b03c4787-0abf-456e-8f11-e9ae79f8c6ab. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1436.453541] env[62814]: DEBUG nova.network.neutron [req-28a9102a-d572-4462-a40f-c01de284f55d req-ca2caccc-d591-4ed7-860e-fb22737b3b9e service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Updating instance_info_cache with network_info: [{"id": "b03c4787-0abf-456e-8f11-e9ae79f8c6ab", "address": "fa:16:3e:0c:bc:bc", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb03c4787-0a", "ovs_interfaceid": "b03c4787-0abf-456e-8f11-e9ae79f8c6ab", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1436.955968] env[62814]: DEBUG oslo_concurrency.lockutils [req-28a9102a-d572-4462-a40f-c01de284f55d req-ca2caccc-d591-4ed7-860e-fb22737b3b9e service nova] Releasing lock "refresh_cache-63b4c501-f3c7-4366-bd3e-83ba769f6f9f" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1468.507116] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1468.507559] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 1468.507666] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1469.011145] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1469.011404] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1469.011581] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1469.011737] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1469.012789] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ce1a6c-cbcc-446f-a2ae-9dfd01e4db92 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.020933] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4361cf-15cb-41df-a4d4-c3a055a56152 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.034525] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb68b656-d66b-48d3-a256-c9a431bc4e85 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.040859] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-920a88bb-5da4-44fc-9187-3465f1239d01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1469.070669] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180554MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1469.070669] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1469.070801] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1470.190084] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 711f9623-4c61-47da-b7e5-15a41487b3b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1470.190365] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 63b4c501-f3c7-4366-bd3e-83ba769f6f9f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1470.190442] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1470.190579] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=149GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1470.228298] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-169356ee-02f0-44d2-bae1-355b9dbed43e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.235733] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78c7a6b6-acc2-4de0-9d40-e6f789532578 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.264527] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-659b66e6-4abe-4c88-87ce-07be48700d97 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.271626] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad2aa26-8569-484f-bc10-481274205a69 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1470.285642] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1470.789328] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1471.294884] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1471.295290] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.224s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1471.295332] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1471.295462] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Cleaning up deleted instances with incomplete migration {{(pid=62814) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11903}} [ 1472.793244] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1472.793567] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1472.793725] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1472.793872] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1472.794025] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1472.937582] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1472.937817] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1473.441017] env[62814]: DEBUG nova.compute.utils [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1473.505993] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1473.943777] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1475.003058] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1475.003058] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1475.003058] env[62814]: INFO nova.compute.manager [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Attaching volume ff7155ca-5c8c-438f-9a1c-e08d7d7830a0 to /dev/sdb [ 1475.032992] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a962551-b0f1-4c98-8512-bf70dc785cef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1475.039959] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5145aa4e-9528-4d55-b2bb-aeb4d88e13f0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1475.052601] env[62814]: DEBUG nova.virt.block_device [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Updating existing volume attachment record: 9a5f434d-b0f3-48d7-a68f-503e9a0dedbd {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1476.506528] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1476.506913] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Cleaning up deleted instances {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11865}} [ 1477.015920] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] There are 26 instances to clean {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11874}} [ 1477.016195] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 1a7fb769-f0bf-4387-a3a8-69b05d18f5c9] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1477.522066] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 54eb2a99-c7ef-4faf-a0d8-bef576081904] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1478.025339] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: ba933282-9711-4e99-ad2d-6ad4c6c516cc] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1478.528730] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: add41cff-326f-46f4-b768-1f08ace2b31f] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1479.032609] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: d043b12d-a316-46ab-a30f-abec0a7963a6] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1479.536400] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 7b82801d-7a0f-4baf-ad2e-566174d3eb35] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1479.594882] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1479.595142] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845891', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'name': 'volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '63b4c501-f3c7-4366-bd3e-83ba769f6f9f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'serial': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1479.596064] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23a1570-5228-4eb2-949b-2a4efc238ee7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.612631] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7993c85-97c9-415c-8174-1437d3c1cb4c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.636875] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0/volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1479.637139] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea5cb75a-131e-4e0f-a0e5-ffa4cb997bde {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1479.654888] env[62814]: DEBUG oslo_vmware.api [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1479.654888] env[62814]: value = "task-4294960" [ 1479.654888] env[62814]: _type = "Task" [ 1479.654888] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1479.663384] env[62814]: DEBUG oslo_vmware.api [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294960, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1480.039490] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: ae9180b8-6caa-44ef-93a8-eb3a7681d224] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1480.163545] env[62814]: DEBUG oslo_vmware.api [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294960, 'name': ReconfigVM_Task, 'duration_secs': 0.322712} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1480.163822] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfigured VM instance instance-0000007d to attach disk [datastore2] volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0/volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1480.168452] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c3ada09-05b6-4ebb-bfd4-307baf75bddd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1480.182423] env[62814]: DEBUG oslo_vmware.api [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1480.182423] env[62814]: value = "task-4294961" [ 1480.182423] env[62814]: _type = "Task" [ 1480.182423] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1480.189565] env[62814]: DEBUG oslo_vmware.api [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294961, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1480.543102] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: b5a2fb6c-7078-4f3c-b511-014beb96391d] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1480.693551] env[62814]: DEBUG oslo_vmware.api [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294961, 'name': ReconfigVM_Task, 'duration_secs': 0.154931} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1480.693855] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845891', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'name': 'volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '63b4c501-f3c7-4366-bd3e-83ba769f6f9f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'serial': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1481.046656] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: fed47b28-64fb-4af9-9f8e-97a63afda514] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1481.550886] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: c9b67326-3058-4fa6-a094-b47bf8663444] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1481.730149] env[62814]: DEBUG nova.objects.instance [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lazy-loading 'flavor' on Instance uuid 63b4c501-f3c7-4366-bd3e-83ba769f6f9f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1482.053817] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 0fdb57c1-7a2a-455d-acb8-9f342ef1dbac] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1482.235467] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4eee1cef-6aab-4b83-8178-d0290d1e009e tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.233s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1482.557411] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 44d94dc4-d7b3-4799-9044-9c1ecc80c88c] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1482.679625] env[62814]: INFO nova.compute.manager [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Rebuilding instance [ 1482.727812] env[62814]: DEBUG nova.compute.manager [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1482.728710] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77b0f67-7b67-4391-a05a-f21833d0092f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.060730] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: bf2ba498-7c47-4985-a84a-475037deec55] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1483.564631] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 40e3c00b-2129-476e-bc67-a0ef13781990] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1483.742326] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1483.742881] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5124f1a2-099c-475d-82c7-fef039d1a7e1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1483.750859] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1483.750859] env[62814]: value = "task-4294962" [ 1483.750859] env[62814]: _type = "Task" [ 1483.750859] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1483.759230] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294962, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1484.068465] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 881e9481-a5a9-489b-8abd-8efdaa239a11] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1484.261055] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294962, 'name': PowerOffVM_Task, 'duration_secs': 0.213671} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1484.261582] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1484.314529] env[62814]: INFO nova.compute.manager [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Detaching volume ff7155ca-5c8c-438f-9a1c-e08d7d7830a0 [ 1484.343129] env[62814]: INFO nova.virt.block_device [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Attempting to driver detach volume ff7155ca-5c8c-438f-9a1c-e08d7d7830a0 from mountpoint /dev/sdb [ 1484.343370] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1484.343557] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845891', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'name': 'volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '63b4c501-f3c7-4366-bd3e-83ba769f6f9f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'serial': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1484.344432] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f460d4-64d6-4934-b1d6-7e20c38f07e3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.365378] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d0b3706-ef8e-46a4-8baa-abbbf73d132a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.371613] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea24d08-7718-449a-b283-0102f1885380 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.390748] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aff5545-4baf-4f34-90b1-1710ef5945f1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.404427] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] The volume has not been displaced from its original location: [datastore2] volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0/volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1484.409530] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfiguring VM instance instance-0000007d to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1484.409775] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-de5cbc1c-a46a-4a1d-be66-8387f3ddee70 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.426503] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1484.426503] env[62814]: value = "task-4294963" [ 1484.426503] env[62814]: _type = "Task" [ 1484.426503] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1484.433475] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294963, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1484.572445] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 53e13866-0174-4867-9fd9-2ea3f5ee399d] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1484.937849] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294963, 'name': ReconfigVM_Task, 'duration_secs': 0.218797} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1484.938079] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfigured VM instance instance-0000007d to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1484.942550] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fa646cfd-570e-45a7-9d38-13b2a9c10671 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1484.956895] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1484.956895] env[62814]: value = "task-4294964" [ 1484.956895] env[62814]: _type = "Task" [ 1484.956895] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1484.964362] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294964, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1485.076234] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 29468dc9-4254-4ac6-989c-c7fa0c4cc682] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1485.467189] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294964, 'name': ReconfigVM_Task, 'duration_secs': 0.141769} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1485.467524] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845891', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'name': 'volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '63b4c501-f3c7-4366-bd3e-83ba769f6f9f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'serial': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1485.580243] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: d65ccde2-b31a-4032-8795-1d609fdfcc73] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1486.083553] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 606e31eb-2349-427f-9c9b-ed9dc5b385f5] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1486.514026] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1486.514221] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d38f587-d4b9-47c3-a869-d7c7718c06c7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.521374] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1486.521374] env[62814]: value = "task-4294965" [ 1486.521374] env[62814]: _type = "Task" [ 1486.521374] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1486.529607] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] VM already powered off {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1486.530050] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1486.530050] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845891', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'name': 'volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '63b4c501-f3c7-4366-bd3e-83ba769f6f9f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'serial': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1486.530722] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981c0c7c-3210-4634-9768-20a98d3c55dc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.547746] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99507a91-45f4-4ee9-b876-10790d73a011 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.553051] env[62814]: WARNING nova.virt.vmwareapi.driver [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 1486.553227] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1486.554049] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cce6026-8aa2-4587-9a70-89f979914e5b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.560024] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1486.560024] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66f5eba2-2514-46b5-b3eb-46ce10c5ee03 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.586626] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 33ab2d9e-e960-40ae-acf9-4fea0d9c830e] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1486.620703] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1486.622032] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1486.622032] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleting the datastore file [datastore2] 63b4c501-f3c7-4366-bd3e-83ba769f6f9f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1486.622032] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-afc45f64-909f-44f9-a46f-d0c2508fd03a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1486.627732] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1486.627732] env[62814]: value = "task-4294967" [ 1486.627732] env[62814]: _type = "Task" [ 1486.627732] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1486.635161] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294967, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1487.091055] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: b1d88997-e52f-41bd-b1b4-dd096d20d60a] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1487.137554] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294967, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.122803} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1487.137811] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1487.137988] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1487.138239] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1487.594812] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 2d7a18f3-456b-470c-a759-632e5d79a8f0] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1487.642745] env[62814]: INFO nova.virt.block_device [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Booting with volume ff7155ca-5c8c-438f-9a1c-e08d7d7830a0 at /dev/sdb [ 1487.671818] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-26dc54b8-efcc-4e9d-b1c6-6256067b7c31 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.683654] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab92f4b-40c2-4916-b387-d90d46947837 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.705244] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-471e5287-4669-4343-9737-550d678d814b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.712617] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e8e65a-3868-4b12-ba99-9387b2ee1665 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.735335] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f306b1d-a0b1-48dc-97fc-a40c8cc141fd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.741104] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62e3259-f092-450a-a305-39cd49c61a13 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1487.753442] env[62814]: DEBUG nova.virt.block_device [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Updating existing volume attachment record: abab0f2e-5fdc-442e-94ec-52d6ba30aec5 {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1488.098168] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: f00d414e-1dfd-43cf-9245-7cbeea8850b3] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1488.601968] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 1f13051e-e51b-4981-9445-d5420c3c1818] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1489.104851] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: edcdb1e5-ed9d-49e6-97e6-ea7629682547] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1489.608194] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: 7350d352-9336-40b8-81a6-0a4795d9f8dd] Instance has had 0 of 5 cleanup attempts {{(pid=62814) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11878}} [ 1489.862221] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1489.862587] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1489.862779] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1489.862969] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1489.863134] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1489.863282] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1489.863493] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1489.863648] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1489.863812] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1489.863973] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1489.864160] env[62814]: DEBUG nova.virt.hardware [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1489.864998] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db134ecc-1486-4379-b833-8246e80b990f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1489.872430] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93126d71-49a2-4e5c-91da-8f79acb7ee35 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1489.885399] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:bc:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ffb921-796a-40fe-9662-d3fc01547dcb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b03c4787-0abf-456e-8f11-e9ae79f8c6ab', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1489.892615] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1489.893141] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1489.893344] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-89893293-e711-4dbc-a055-230e7f8133ca {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1489.912159] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1489.912159] env[62814]: value = "task-4294968" [ 1489.912159] env[62814]: _type = "Task" [ 1489.912159] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1489.922964] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294968, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1490.111821] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1490.426233] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294968, 'name': CreateVM_Task, 'duration_secs': 0.287616} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1490.426439] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1490.427326] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1490.427538] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1490.427967] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1490.428310] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-595d5c49-a66a-427d-b30d-da09160ca28a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.433526] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1490.433526] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d235f1-c959-1664-0c9c-e46142bf8645" [ 1490.433526] env[62814]: _type = "Task" [ 1490.433526] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.443449] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d235f1-c959-1664-0c9c-e46142bf8645, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1490.943793] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52d235f1-c959-1664-0c9c-e46142bf8645, 'name': SearchDatastore_Task, 'duration_secs': 0.009104} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1490.944223] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1490.944345] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1490.944577] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1490.944725] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1490.944915] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1490.945191] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9cca052a-9c10-4df5-bf55-1d583652d27b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.953019] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1490.953191] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1490.953844] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d71b833-93c9-4fae-b3fc-0641bdda2e37 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1490.958326] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1490.958326] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52066c92-8811-d29b-802f-6bfffd191c6c" [ 1490.958326] env[62814]: _type = "Task" [ 1490.958326] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1490.966127] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52066c92-8811-d29b-802f-6bfffd191c6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.468758] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52066c92-8811-d29b-802f-6bfffd191c6c, 'name': SearchDatastore_Task, 'duration_secs': 0.008884} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1491.469496] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62fe7915-b949-4484-b84c-be650fc4cf1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.474873] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1491.474873] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c270fd-1505-543f-4173-b15c7bfc0ea3" [ 1491.474873] env[62814]: _type = "Task" [ 1491.474873] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1491.481922] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c270fd-1505-543f-4173-b15c7bfc0ea3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1491.985024] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52c270fd-1505-543f-4173-b15c7bfc0ea3, 'name': SearchDatastore_Task, 'duration_secs': 0.009455} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1491.985484] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1491.985569] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 63b4c501-f3c7-4366-bd3e-83ba769f6f9f/63b4c501-f3c7-4366-bd3e-83ba769f6f9f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1491.985820] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d699b59-7836-4e1d-b883-c9e83b49f4d7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1491.991942] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1491.991942] env[62814]: value = "task-4294969" [ 1491.991942] env[62814]: _type = "Task" [ 1491.991942] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1491.998891] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294969, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1492.503502] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294969, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.426565} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1492.503767] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 63b4c501-f3c7-4366-bd3e-83ba769f6f9f/63b4c501-f3c7-4366-bd3e-83ba769f6f9f.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1492.503975] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1492.504234] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01b4d08e-1a8d-478a-b9db-c595cc357d1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1492.511112] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1492.511112] env[62814]: value = "task-4294970" [ 1492.511112] env[62814]: _type = "Task" [ 1492.511112] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1492.519093] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294970, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1493.020931] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294970, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054584} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1493.021277] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1493.022025] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1662ba9-7d64-4532-9f10-26cc7defd453 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.043683] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] 63b4c501-f3c7-4366-bd3e-83ba769f6f9f/63b4c501-f3c7-4366-bd3e-83ba769f6f9f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1493.045051] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f96f1996-8204-4894-bb54-dc4978863bea {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.062394] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1493.062394] env[62814]: value = "task-4294971" [ 1493.062394] env[62814]: _type = "Task" [ 1493.062394] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1493.069909] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294971, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1493.572235] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294971, 'name': ReconfigVM_Task, 'duration_secs': 0.261057} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1493.572695] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfigured VM instance instance-0000007d to attach disk [datastore2] 63b4c501-f3c7-4366-bd3e-83ba769f6f9f/63b4c501-f3c7-4366-bd3e-83ba769f6f9f.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1493.573987] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_secret_uuid': None, 'size': 0, 'device_type': 'disk', 'device_name': '/dev/sda', 'boot_index': 0, 'encrypted': False, 'guest_format': None, 'encryption_format': None, 'disk_bus': None, 'encryption_options': None, 'image_id': 'e7d68632-de75-4206-8f75-4abd879e1d22'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'abab0f2e-5fdc-442e-94ec-52d6ba30aec5', 'mount_device': '/dev/sdb', 'device_type': None, 'delete_on_termination': False, 'boot_index': None, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845891', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'name': 'volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '63b4c501-f3c7-4366-bd3e-83ba769f6f9f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'serial': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0'}, 'guest_format': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=62814) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1493.574396] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1493.575024] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845891', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'name': 'volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '63b4c501-f3c7-4366-bd3e-83ba769f6f9f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'serial': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1493.575493] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6708418e-9700-4fa4-be1c-ef0f374260c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.590554] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e94b7700-f6e9-4bfa-a7c4-eb35e38363ba {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.614500] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfiguring VM instance instance-0000007d to attach disk [datastore2] volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0/volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1493.615182] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1d21981-6eb5-4c61-8ce1-65594675134f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1493.632756] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1493.632756] env[62814]: value = "task-4294972" [ 1493.632756] env[62814]: _type = "Task" [ 1493.632756] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1493.640324] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294972, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1494.144054] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294972, 'name': ReconfigVM_Task, 'duration_secs': 0.28776} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1494.144054] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfigured VM instance instance-0000007d to attach disk [datastore2] volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0/volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1494.148071] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d58ed81b-cf8b-4c17-b182-73849cf9f43f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.162617] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1494.162617] env[62814]: value = "task-4294973" [ 1494.162617] env[62814]: _type = "Task" [ 1494.162617] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1494.169894] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294973, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1494.673167] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294973, 'name': ReconfigVM_Task, 'duration_secs': 0.138737} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1494.673482] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845891', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'name': 'volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '63b4c501-f3c7-4366-bd3e-83ba769f6f9f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'serial': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1494.674072] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7032e039-02ff-406b-8262-b0ea79686d8e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1494.680776] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1494.680776] env[62814]: value = "task-4294974" [ 1494.680776] env[62814]: _type = "Task" [ 1494.680776] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1494.689356] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294974, 'name': Rename_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1495.190585] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294974, 'name': Rename_Task, 'duration_secs': 0.139791} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1495.190941] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1495.191248] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f7da0163-35db-45a9-957d-e4b20d54dfc2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1495.198275] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1495.198275] env[62814]: value = "task-4294975" [ 1495.198275] env[62814]: _type = "Task" [ 1495.198275] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1495.208657] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294975, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1495.708755] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294975, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1496.207687] env[62814]: DEBUG oslo_vmware.api [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294975, 'name': PowerOnVM_Task, 'duration_secs': 0.549501} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1496.208131] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1496.208200] env[62814]: DEBUG nova.compute.manager [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1496.208942] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19370c2d-bc03-45af-be0a-db03fdf8654d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1496.726672] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1496.726912] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1496.727105] env[62814]: DEBUG nova.objects.instance [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62814) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1497.736058] env[62814]: DEBUG oslo_concurrency.lockutils [None req-9ffea827-49a6-45d8-a143-be02b82d2e39 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1533.748287] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1533.748652] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1534.251536] env[62814]: INFO nova.compute.manager [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Detaching volume ff7155ca-5c8c-438f-9a1c-e08d7d7830a0 [ 1534.315121] env[62814]: INFO nova.virt.block_device [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Attempting to driver detach volume ff7155ca-5c8c-438f-9a1c-e08d7d7830a0 from mountpoint /dev/sdb [ 1534.315392] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1534.315587] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845891', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'name': 'volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '63b4c501-f3c7-4366-bd3e-83ba769f6f9f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'serial': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1534.316552] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6db0b61-402e-4d79-b996-e01346f7698e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.338272] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec20b0b-564a-4a95-ba60-6de773fe6fd8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.344977] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e286d008-5231-4504-acbf-909dcff26811 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.364478] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf7eb05e-4e76-46ef-ad54-237fc049e8a2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.378843] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] The volume has not been displaced from its original location: [datastore2] volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0/volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1534.383863] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfiguring VM instance instance-0000007d to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1534.384136] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0e06b90-fb94-40be-836e-7b667a95f5bc {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.401221] env[62814]: DEBUG oslo_vmware.api [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1534.401221] env[62814]: value = "task-4294976" [ 1534.401221] env[62814]: _type = "Task" [ 1534.401221] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1534.408475] env[62814]: DEBUG oslo_vmware.api [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294976, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1534.910411] env[62814]: DEBUG oslo_vmware.api [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294976, 'name': ReconfigVM_Task, 'duration_secs': 0.21302} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1534.910788] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Reconfigured VM instance instance-0000007d to detach disk 2001 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1534.915145] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fcd00afc-16c4-4ef0-a390-e56744f5edd1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1534.930110] env[62814]: DEBUG oslo_vmware.api [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1534.930110] env[62814]: value = "task-4294977" [ 1534.930110] env[62814]: _type = "Task" [ 1534.930110] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1534.937399] env[62814]: DEBUG oslo_vmware.api [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294977, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1535.439684] env[62814]: DEBUG oslo_vmware.api [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294977, 'name': ReconfigVM_Task, 'duration_secs': 0.133099} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1535.439993] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845891', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'name': 'volume-ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '63b4c501-f3c7-4366-bd3e-83ba769f6f9f', 'attached_at': '', 'detached_at': '', 'volume_id': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0', 'serial': 'ff7155ca-5c8c-438f-9a1c-e08d7d7830a0'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1535.979568] env[62814]: DEBUG nova.objects.instance [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lazy-loading 'flavor' on Instance uuid 63b4c501-f3c7-4366-bd3e-83ba769f6f9f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1536.988599] env[62814]: DEBUG oslo_concurrency.lockutils [None req-1f5f81bb-023b-41d5-8ba7-feb222d26dc5 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.240s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1537.987605] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1537.987917] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1537.988152] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1537.988348] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1537.988506] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1537.990688] env[62814]: INFO nova.compute.manager [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Terminating instance [ 1538.494239] env[62814]: DEBUG nova.compute.manager [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1538.494563] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1538.495505] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d680618a-ddca-4d31-9c79-89f0c6d4a07b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.503337] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1538.503544] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-88658088-12a3-4eb5-8f4b-b49621a6fdd5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1538.509396] env[62814]: DEBUG oslo_vmware.api [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1538.509396] env[62814]: value = "task-4294978" [ 1538.509396] env[62814]: _type = "Task" [ 1538.509396] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1538.517279] env[62814]: DEBUG oslo_vmware.api [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1539.019705] env[62814]: DEBUG oslo_vmware.api [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294978, 'name': PowerOffVM_Task, 'duration_secs': 0.145339} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1539.020078] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1539.020151] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1539.020389] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71f28d36-9bf9-41fd-a816-f68e6fc6565d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1539.078571] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1539.078804] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1539.078983] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleting the datastore file [datastore2] 63b4c501-f3c7-4366-bd3e-83ba769f6f9f {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1539.079250] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef13da48-851a-4e92-bd7d-0dfd0a657e68 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1539.086293] env[62814]: DEBUG oslo_vmware.api [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1539.086293] env[62814]: value = "task-4294980" [ 1539.086293] env[62814]: _type = "Task" [ 1539.086293] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1539.093313] env[62814]: DEBUG oslo_vmware.api [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1539.596361] env[62814]: DEBUG oslo_vmware.api [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294980, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133888} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1539.596602] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1539.596784] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1539.596958] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1539.597151] env[62814]: INFO nova.compute.manager [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1539.597387] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1539.597570] env[62814]: DEBUG nova.compute.manager [-] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1539.597663] env[62814]: DEBUG nova.network.neutron [-] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1540.140492] env[62814]: DEBUG nova.compute.manager [req-c990b769-0bf9-4688-a162-366770805e31 req-8c277afc-4255-484e-b3d9-4102afbfd861 service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Received event network-vif-deleted-b03c4787-0abf-456e-8f11-e9ae79f8c6ab {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1540.140723] env[62814]: INFO nova.compute.manager [req-c990b769-0bf9-4688-a162-366770805e31 req-8c277afc-4255-484e-b3d9-4102afbfd861 service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Neutron deleted interface b03c4787-0abf-456e-8f11-e9ae79f8c6ab; detaching it from the instance and deleting it from the info cache [ 1540.140814] env[62814]: DEBUG nova.network.neutron [req-c990b769-0bf9-4688-a162-366770805e31 req-8c277afc-4255-484e-b3d9-4102afbfd861 service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1540.585858] env[62814]: DEBUG nova.network.neutron [-] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1540.643335] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e27a2358-bbe8-4d42-9528-5ba40604ff4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.654453] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-571a919f-705f-4526-9a03-46220627c96f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1540.679672] env[62814]: DEBUG nova.compute.manager [req-c990b769-0bf9-4688-a162-366770805e31 req-8c277afc-4255-484e-b3d9-4102afbfd861 service nova] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Detach interface failed, port_id=b03c4787-0abf-456e-8f11-e9ae79f8c6ab, reason: Instance 63b4c501-f3c7-4366-bd3e-83ba769f6f9f could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1541.088544] env[62814]: INFO nova.compute.manager [-] [instance: 63b4c501-f3c7-4366-bd3e-83ba769f6f9f] Took 1.49 seconds to deallocate network for instance. [ 1541.595486] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1541.595877] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1541.596085] env[62814]: DEBUG nova.objects.instance [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lazy-loading 'resources' on Instance uuid 63b4c501-f3c7-4366-bd3e-83ba769f6f9f {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1542.118139] env[62814]: DEBUG nova.scheduler.client.report [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Refreshing inventories for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 1542.133717] env[62814]: DEBUG nova.scheduler.client.report [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Updating ProviderTree inventory for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 1542.133945] env[62814]: DEBUG nova.compute.provider_tree [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Updating inventory in ProviderTree for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1542.145600] env[62814]: DEBUG nova.scheduler.client.report [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Refreshing aggregate associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, aggregates: None {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 1542.166625] env[62814]: DEBUG nova.scheduler.client.report [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Refreshing trait associations for resource provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NODE,HW_ARCH_X86_64 {{(pid=62814) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 1542.208603] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015680e9-2c78-4df3-8f09-559863a34dff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.216361] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-977446f7-5dc0-4f59-94f2-42c089435691 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.245751] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934c0005-daa7-4347-b8c8-ad91ebe49ee8 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.255473] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44309b87-b366-43d2-9de9-3b5bcfbbf283 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1542.268985] env[62814]: DEBUG nova.compute.provider_tree [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1542.614930] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1542.615310] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1542.772351] env[62814]: DEBUG nova.scheduler.client.report [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1543.122065] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1543.122262] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1543.122262] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1543.122408] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1543.122554] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1543.122697] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1543.122830] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 1543.122974] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1543.277275] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.681s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1543.304829] env[62814]: INFO nova.scheduler.client.report [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted allocations for instance 63b4c501-f3c7-4366-bd3e-83ba769f6f9f [ 1543.626152] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1543.626629] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1543.626629] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1543.626759] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1543.627629] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e406d423-da12-4a22-96f9-083fb6e6b2c5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.635936] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bea110-ee03-4b34-a28f-472681aa1b4f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.650418] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9008cf9d-642f-409f-ac4c-b61969687bc1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.656970] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e671674-009b-4346-8d8c-26b5f0542700 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1543.686189] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180383MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1543.686335] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1543.686559] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1543.812800] env[62814]: DEBUG oslo_concurrency.lockutils [None req-d1f1a2f4-29c4-4636-9609-c75bd5d0ab40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "63b4c501-f3c7-4366-bd3e-83ba769f6f9f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.825s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1544.711939] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 711f9623-4c61-47da-b7e5-15a41487b3b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1544.712233] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1544.712362] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=149GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1544.739484] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dc0a382-bdf4-414d-b5ec-eb6b784b1e42 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1544.747564] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c443078e-8e96-4bc6-b242-bbc8bd611007 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1544.776655] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fba45fa-e13d-44e0-a663-4f8e05372628 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1544.783593] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb219cf-ea59-4890-941a-6bab0b38212e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1544.796394] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1545.300171] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1545.301525] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1545.301714] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.615s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1545.506796] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "7099d3d1-c165-4600-b4ab-610161a9834d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1545.507086] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "7099d3d1-c165-4600-b4ab-610161a9834d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1546.009277] env[62814]: DEBUG nova.compute.manager [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1546.530682] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1546.530960] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1546.532378] env[62814]: INFO nova.compute.claims [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1547.778940] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae6652e-7402-404c-8f51-9ced7d319aaf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.786264] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de868fc1-66e9-48bb-b16e-6756d33e3452 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.814555] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90d77e5-cd24-4b93-bae3-8aacafc0f364 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.821007] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566be1db-075e-47e7-b049-2b1e84171da5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1547.833358] env[62814]: DEBUG nova.compute.provider_tree [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1548.336871] env[62814]: DEBUG nova.scheduler.client.report [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1548.841789] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1548.842336] env[62814]: DEBUG nova.compute.manager [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1549.347151] env[62814]: DEBUG nova.compute.utils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1549.348542] env[62814]: DEBUG nova.compute.manager [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1549.348709] env[62814]: DEBUG nova.network.neutron [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1549.395519] env[62814]: DEBUG nova.policy [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9516bbf4242140059eeb2182d3e7a58b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb7d25c2161f40db8197b1a56517d7f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1549.619197] env[62814]: DEBUG nova.network.neutron [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Successfully created port: 9489e63a-3e31-4600-b530-7bde469fca61 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1549.852326] env[62814]: DEBUG nova.compute.manager [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1550.861585] env[62814]: DEBUG nova.compute.manager [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1550.887664] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-06-21T05:32:20Z,direct_url=,disk_format='vmdk',id=e7d68632-de75-4206-8f75-4abd879e1d22,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='32bba7a718ae4b3ba6a475c985d52bd7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-06-21T05:32:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1550.887919] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1550.888111] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1550.888301] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1550.888445] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1550.888588] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1550.888790] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1550.888968] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1550.889151] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1550.889312] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1550.889478] env[62814]: DEBUG nova.virt.hardware [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1550.890330] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4059677-642d-49a1-a9c7-90e914b92fee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1550.898433] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1f2bf4-5523-4fc9-95e0-1223a88e12ee {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1551.019570] env[62814]: DEBUG nova.compute.manager [req-5cb8624d-6977-4e3b-a8cf-110e1c7b7816 req-5521a72e-c2bc-4549-82c4-17db48e0b6e4 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Received event network-vif-plugged-9489e63a-3e31-4600-b530-7bde469fca61 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1551.019927] env[62814]: DEBUG oslo_concurrency.lockutils [req-5cb8624d-6977-4e3b-a8cf-110e1c7b7816 req-5521a72e-c2bc-4549-82c4-17db48e0b6e4 service nova] Acquiring lock "7099d3d1-c165-4600-b4ab-610161a9834d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1551.020281] env[62814]: DEBUG oslo_concurrency.lockutils [req-5cb8624d-6977-4e3b-a8cf-110e1c7b7816 req-5521a72e-c2bc-4549-82c4-17db48e0b6e4 service nova] Lock "7099d3d1-c165-4600-b4ab-610161a9834d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1551.020546] env[62814]: DEBUG oslo_concurrency.lockutils [req-5cb8624d-6977-4e3b-a8cf-110e1c7b7816 req-5521a72e-c2bc-4549-82c4-17db48e0b6e4 service nova] Lock "7099d3d1-c165-4600-b4ab-610161a9834d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1551.020814] env[62814]: DEBUG nova.compute.manager [req-5cb8624d-6977-4e3b-a8cf-110e1c7b7816 req-5521a72e-c2bc-4549-82c4-17db48e0b6e4 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] No waiting events found dispatching network-vif-plugged-9489e63a-3e31-4600-b530-7bde469fca61 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1551.021115] env[62814]: WARNING nova.compute.manager [req-5cb8624d-6977-4e3b-a8cf-110e1c7b7816 req-5521a72e-c2bc-4549-82c4-17db48e0b6e4 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Received unexpected event network-vif-plugged-9489e63a-3e31-4600-b530-7bde469fca61 for instance with vm_state building and task_state spawning. [ 1551.098407] env[62814]: DEBUG nova.network.neutron [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Successfully updated port: 9489e63a-3e31-4600-b530-7bde469fca61 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1551.602981] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "refresh_cache-7099d3d1-c165-4600-b4ab-610161a9834d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1551.603098] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "refresh_cache-7099d3d1-c165-4600-b4ab-610161a9834d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1551.603257] env[62814]: DEBUG nova.network.neutron [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1552.146384] env[62814]: DEBUG nova.network.neutron [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1552.270788] env[62814]: DEBUG nova.network.neutron [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Updating instance_info_cache with network_info: [{"id": "9489e63a-3e31-4600-b530-7bde469fca61", "address": "fa:16:3e:90:d8:7d", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9489e63a-3e", "ovs_interfaceid": "9489e63a-3e31-4600-b530-7bde469fca61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1552.773919] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "refresh_cache-7099d3d1-c165-4600-b4ab-610161a9834d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1552.774381] env[62814]: DEBUG nova.compute.manager [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Instance network_info: |[{"id": "9489e63a-3e31-4600-b530-7bde469fca61", "address": "fa:16:3e:90:d8:7d", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9489e63a-3e", "ovs_interfaceid": "9489e63a-3e31-4600-b530-7bde469fca61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1552.774896] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:d8:7d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ffb921-796a-40fe-9662-d3fc01547dcb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9489e63a-3e31-4600-b530-7bde469fca61', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1552.782590] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1552.782893] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1552.783513] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-da37156a-a2b4-4f86-905d-79fc24178888 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1552.805191] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1552.805191] env[62814]: value = "task-4294981" [ 1552.805191] env[62814]: _type = "Task" [ 1552.805191] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1552.812771] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294981, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1553.051202] env[62814]: DEBUG nova.compute.manager [req-9557f7f2-eb58-471c-aaac-0719f5151b7b req-243257e2-e987-492d-a66b-ee43fff71758 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Received event network-changed-9489e63a-3e31-4600-b530-7bde469fca61 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1553.051468] env[62814]: DEBUG nova.compute.manager [req-9557f7f2-eb58-471c-aaac-0719f5151b7b req-243257e2-e987-492d-a66b-ee43fff71758 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Refreshing instance network info cache due to event network-changed-9489e63a-3e31-4600-b530-7bde469fca61. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1553.051805] env[62814]: DEBUG oslo_concurrency.lockutils [req-9557f7f2-eb58-471c-aaac-0719f5151b7b req-243257e2-e987-492d-a66b-ee43fff71758 service nova] Acquiring lock "refresh_cache-7099d3d1-c165-4600-b4ab-610161a9834d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1553.052072] env[62814]: DEBUG oslo_concurrency.lockutils [req-9557f7f2-eb58-471c-aaac-0719f5151b7b req-243257e2-e987-492d-a66b-ee43fff71758 service nova] Acquired lock "refresh_cache-7099d3d1-c165-4600-b4ab-610161a9834d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1553.052283] env[62814]: DEBUG nova.network.neutron [req-9557f7f2-eb58-471c-aaac-0719f5151b7b req-243257e2-e987-492d-a66b-ee43fff71758 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Refreshing network info cache for port 9489e63a-3e31-4600-b530-7bde469fca61 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1553.315561] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294981, 'name': CreateVM_Task, 'duration_secs': 0.289488} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1553.315923] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1553.316363] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1553.316528] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1553.316853] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 1553.317108] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-97b836e7-4b3b-4662-88cb-07a96d232344 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1553.321972] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1553.321972] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a85211-d7ff-7205-17b5-601a765d2f01" [ 1553.321972] env[62814]: _type = "Task" [ 1553.321972] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1553.329313] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a85211-d7ff-7205-17b5-601a765d2f01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1553.781768] env[62814]: DEBUG nova.network.neutron [req-9557f7f2-eb58-471c-aaac-0719f5151b7b req-243257e2-e987-492d-a66b-ee43fff71758 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Updated VIF entry in instance network info cache for port 9489e63a-3e31-4600-b530-7bde469fca61. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1553.782147] env[62814]: DEBUG nova.network.neutron [req-9557f7f2-eb58-471c-aaac-0719f5151b7b req-243257e2-e987-492d-a66b-ee43fff71758 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Updating instance_info_cache with network_info: [{"id": "9489e63a-3e31-4600-b530-7bde469fca61", "address": "fa:16:3e:90:d8:7d", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9489e63a-3e", "ovs_interfaceid": "9489e63a-3e31-4600-b530-7bde469fca61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1553.832841] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52a85211-d7ff-7205-17b5-601a765d2f01, 'name': SearchDatastore_Task, 'duration_secs': 0.011374} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1553.833152] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1553.833375] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Processing image e7d68632-de75-4206-8f75-4abd879e1d22 {{(pid=62814) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1553.833605] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1553.833746] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1553.833918] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1553.834197] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c0af0a0c-b5ef-439e-ae38-61980f85c7d1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1553.842710] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62814) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1553.842824] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62814) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1553.843502] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7373566e-bb1d-48d7-9400-95bd1d7894b4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1553.848760] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1553.848760] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5224b572-f327-cc56-407d-e89c05ca1732" [ 1553.848760] env[62814]: _type = "Task" [ 1553.848760] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1553.855524] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5224b572-f327-cc56-407d-e89c05ca1732, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1554.285467] env[62814]: DEBUG oslo_concurrency.lockutils [req-9557f7f2-eb58-471c-aaac-0719f5151b7b req-243257e2-e987-492d-a66b-ee43fff71758 service nova] Releasing lock "refresh_cache-7099d3d1-c165-4600-b4ab-610161a9834d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1554.358841] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5224b572-f327-cc56-407d-e89c05ca1732, 'name': SearchDatastore_Task, 'duration_secs': 0.010088} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1554.359594] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74cbb7b7-3039-4e82-aa31-6a0e4f1c2f32 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1554.364153] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1554.364153] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5233c38c-35f3-10e7-9e25-8c3e4cd05c33" [ 1554.364153] env[62814]: _type = "Task" [ 1554.364153] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1554.370943] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5233c38c-35f3-10e7-9e25-8c3e4cd05c33, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1554.873802] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]5233c38c-35f3-10e7-9e25-8c3e4cd05c33, 'name': SearchDatastore_Task, 'duration_secs': 0.009932} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1554.874067] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "[datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1554.874322] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 7099d3d1-c165-4600-b4ab-610161a9834d/7099d3d1-c165-4600-b4ab-610161a9834d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1554.874560] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90459013-883c-4e35-b779-4d3933e4ac1a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1554.881012] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1554.881012] env[62814]: value = "task-4294982" [ 1554.881012] env[62814]: _type = "Task" [ 1554.881012] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1554.888010] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294982, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1555.390953] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294982, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1555.892266] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294982, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546186} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1555.892562] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/e7d68632-de75-4206-8f75-4abd879e1d22/e7d68632-de75-4206-8f75-4abd879e1d22.vmdk to [datastore2] 7099d3d1-c165-4600-b4ab-610161a9834d/7099d3d1-c165-4600-b4ab-610161a9834d.vmdk {{(pid=62814) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1555.892801] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Extending root virtual disk to 1048576 {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1555.893087] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae16791c-86ad-4af7-a48c-ff91d79a86c9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1555.899313] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1555.899313] env[62814]: value = "task-4294983" [ 1555.899313] env[62814]: _type = "Task" [ 1555.899313] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1555.906285] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294983, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1556.408686] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294983, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065846} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1556.409064] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Extended root virtual disk {{(pid=62814) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1556.409617] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624ce020-009e-42ee-9ca4-94faf4dfda01 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1556.432023] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Reconfiguring VM instance instance-0000007e to attach disk [datastore2] 7099d3d1-c165-4600-b4ab-610161a9834d/7099d3d1-c165-4600-b4ab-610161a9834d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1556.432287] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72941698-f72e-4a25-854d-4123f1c0ffda {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1556.451824] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1556.451824] env[62814]: value = "task-4294984" [ 1556.451824] env[62814]: _type = "Task" [ 1556.451824] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1556.459123] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294984, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1556.961879] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294984, 'name': ReconfigVM_Task, 'duration_secs': 0.272559} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1556.962237] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Reconfigured VM instance instance-0000007e to attach disk [datastore2] 7099d3d1-c165-4600-b4ab-610161a9834d/7099d3d1-c165-4600-b4ab-610161a9834d.vmdk or device None with type sparse {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1556.962923] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a51062e3-3671-4a0d-8118-f7f83203377f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1556.970189] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1556.970189] env[62814]: value = "task-4294985" [ 1556.970189] env[62814]: _type = "Task" [ 1556.970189] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1556.977457] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294985, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1557.478940] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294985, 'name': Rename_Task, 'duration_secs': 0.148156} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1557.479286] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1557.479553] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0245ff1-4412-4eee-a6bf-ff25e4e87dff {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1557.485760] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1557.485760] env[62814]: value = "task-4294986" [ 1557.485760] env[62814]: _type = "Task" [ 1557.485760] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1557.492771] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294986, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1557.998546] env[62814]: DEBUG oslo_vmware.api [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294986, 'name': PowerOnVM_Task, 'duration_secs': 0.437682} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1557.998915] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1557.999227] env[62814]: INFO nova.compute.manager [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Took 7.14 seconds to spawn the instance on the hypervisor. [ 1557.999485] env[62814]: DEBUG nova.compute.manager [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1558.000535] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff75084c-50f8-4f70-b840-3b9d75b5effa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1558.522139] env[62814]: INFO nova.compute.manager [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Took 12.01 seconds to build instance. [ 1559.024087] env[62814]: DEBUG oslo_concurrency.lockutils [None req-3ecbfb29-f511-46e4-a32c-642ff8b1d3d1 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "7099d3d1-c165-4600-b4ab-610161a9834d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.517s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1559.567282] env[62814]: DEBUG nova.compute.manager [req-aaeaae69-8256-4e29-a41f-a869c77ef69e req-1f5eabcc-3327-4c0c-bf52-04cccb6deb6e service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Received event network-changed-9489e63a-3e31-4600-b530-7bde469fca61 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1559.567282] env[62814]: DEBUG nova.compute.manager [req-aaeaae69-8256-4e29-a41f-a869c77ef69e req-1f5eabcc-3327-4c0c-bf52-04cccb6deb6e service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Refreshing instance network info cache due to event network-changed-9489e63a-3e31-4600-b530-7bde469fca61. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1559.567588] env[62814]: DEBUG oslo_concurrency.lockutils [req-aaeaae69-8256-4e29-a41f-a869c77ef69e req-1f5eabcc-3327-4c0c-bf52-04cccb6deb6e service nova] Acquiring lock "refresh_cache-7099d3d1-c165-4600-b4ab-610161a9834d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1559.567588] env[62814]: DEBUG oslo_concurrency.lockutils [req-aaeaae69-8256-4e29-a41f-a869c77ef69e req-1f5eabcc-3327-4c0c-bf52-04cccb6deb6e service nova] Acquired lock "refresh_cache-7099d3d1-c165-4600-b4ab-610161a9834d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1559.567731] env[62814]: DEBUG nova.network.neutron [req-aaeaae69-8256-4e29-a41f-a869c77ef69e req-1f5eabcc-3327-4c0c-bf52-04cccb6deb6e service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Refreshing network info cache for port 9489e63a-3e31-4600-b530-7bde469fca61 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1560.285065] env[62814]: DEBUG nova.network.neutron [req-aaeaae69-8256-4e29-a41f-a869c77ef69e req-1f5eabcc-3327-4c0c-bf52-04cccb6deb6e service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Updated VIF entry in instance network info cache for port 9489e63a-3e31-4600-b530-7bde469fca61. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1560.285065] env[62814]: DEBUG nova.network.neutron [req-aaeaae69-8256-4e29-a41f-a869c77ef69e req-1f5eabcc-3327-4c0c-bf52-04cccb6deb6e service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Updating instance_info_cache with network_info: [{"id": "9489e63a-3e31-4600-b530-7bde469fca61", "address": "fa:16:3e:90:d8:7d", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9489e63a-3e", "ovs_interfaceid": "9489e63a-3e31-4600-b530-7bde469fca61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1560.787322] env[62814]: DEBUG oslo_concurrency.lockutils [req-aaeaae69-8256-4e29-a41f-a869c77ef69e req-1f5eabcc-3327-4c0c-bf52-04cccb6deb6e service nova] Releasing lock "refresh_cache-7099d3d1-c165-4600-b4ab-610161a9834d" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1578.669458] env[62814]: DEBUG oslo_concurrency.lockutils [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1578.669880] env[62814]: DEBUG oslo_concurrency.lockutils [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1579.172778] env[62814]: DEBUG nova.compute.manager [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Starting instance... {{(pid=62814) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 1579.694256] env[62814]: DEBUG oslo_concurrency.lockutils [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1579.694542] env[62814]: DEBUG oslo_concurrency.lockutils [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1579.696403] env[62814]: INFO nova.compute.claims [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1580.755119] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8087c7-88ee-4bda-bff4-2f377555af5d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1580.763533] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28f6530a-1f1e-457d-8742-d35301971934 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1580.793226] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25eca293-a42a-4f44-baa2-0f7575d4b442 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1580.800839] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607bd7d1-2f4f-4db8-991f-8f447ba2d6cf {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1580.814744] env[62814]: DEBUG nova.compute.provider_tree [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1581.318325] env[62814]: DEBUG nova.scheduler.client.report [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1581.823871] env[62814]: DEBUG oslo_concurrency.lockutils [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.129s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1581.824347] env[62814]: DEBUG nova.compute.manager [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Start building networks asynchronously for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2836}} [ 1582.330145] env[62814]: DEBUG nova.compute.utils [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Using /dev/sd instead of None {{(pid=62814) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 1582.331685] env[62814]: DEBUG nova.compute.manager [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Allocating IP information in the background. {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1988}} [ 1582.331877] env[62814]: DEBUG nova.network.neutron [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] allocate_for_instance() {{(pid=62814) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1582.385133] env[62814]: DEBUG nova.policy [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9516bbf4242140059eeb2182d3e7a58b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'eb7d25c2161f40db8197b1a56517d7f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62814) authorize /opt/stack/nova/nova/policy.py:192}} [ 1582.632939] env[62814]: DEBUG nova.network.neutron [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Successfully created port: c0fba9cd-6bc1-41dd-902b-4e5546433a59 {{(pid=62814) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1582.835283] env[62814]: DEBUG nova.compute.manager [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Start building block device mappings for instance. {{(pid=62814) _build_resources /opt/stack/nova/nova/compute/manager.py:2871}} [ 1583.341090] env[62814]: INFO nova.virt.block_device [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Booting with volume 16230161-766e-474b-9d7b-4bada573fe8b at /dev/sda [ 1583.381258] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43c654e0-dc2f-4969-b34a-bd6cd919b2a6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1583.391958] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18adb003-c7ad-46fe-a8f3-2cbf8984be17 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1583.420181] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3133b516-1e9a-4762-89d6-9ffe9e207b22 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1583.429646] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52fc15cb-0672-4bc3-815b-b7f22a82fa65 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1583.456541] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82136be-be0c-46ef-8f5d-3926bde72c99 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1583.462953] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c06f9882-36b9-40b3-931a-bc41f799e088 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1583.478283] env[62814]: DEBUG nova.virt.block_device [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating existing volume attachment record: 0761804a-6ad8-4c8f-9444-34ba8ba1a4a5 {{(pid=62814) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1584.022714] env[62814]: DEBUG nova.compute.manager [req-6166abbe-ce05-4d99-99c9-117a04453e1d req-bbb9f926-cc93-4275-8ea0-455543490720 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Received event network-vif-plugged-c0fba9cd-6bc1-41dd-902b-4e5546433a59 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1584.022968] env[62814]: DEBUG oslo_concurrency.lockutils [req-6166abbe-ce05-4d99-99c9-117a04453e1d req-bbb9f926-cc93-4275-8ea0-455543490720 service nova] Acquiring lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1584.023191] env[62814]: DEBUG oslo_concurrency.lockutils [req-6166abbe-ce05-4d99-99c9-117a04453e1d req-bbb9f926-cc93-4275-8ea0-455543490720 service nova] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1584.023361] env[62814]: DEBUG oslo_concurrency.lockutils [req-6166abbe-ce05-4d99-99c9-117a04453e1d req-bbb9f926-cc93-4275-8ea0-455543490720 service nova] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1584.023592] env[62814]: DEBUG nova.compute.manager [req-6166abbe-ce05-4d99-99c9-117a04453e1d req-bbb9f926-cc93-4275-8ea0-455543490720 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] No waiting events found dispatching network-vif-plugged-c0fba9cd-6bc1-41dd-902b-4e5546433a59 {{(pid=62814) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1584.023827] env[62814]: WARNING nova.compute.manager [req-6166abbe-ce05-4d99-99c9-117a04453e1d req-bbb9f926-cc93-4275-8ea0-455543490720 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Received unexpected event network-vif-plugged-c0fba9cd-6bc1-41dd-902b-4e5546433a59 for instance with vm_state building and task_state block_device_mapping. [ 1584.098267] env[62814]: DEBUG nova.network.neutron [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Successfully updated port: c0fba9cd-6bc1-41dd-902b-4e5546433a59 {{(pid=62814) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1584.601181] env[62814]: DEBUG oslo_concurrency.lockutils [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1584.601311] env[62814]: DEBUG oslo_concurrency.lockutils [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1584.601451] env[62814]: DEBUG nova.network.neutron [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1585.142399] env[62814]: DEBUG nova.network.neutron [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Instance cache missing network info. {{(pid=62814) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1585.273570] env[62814]: DEBUG nova.network.neutron [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance_info_cache with network_info: [{"id": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "address": "fa:16:3e:1e:b4:10", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0fba9cd-6b", "ovs_interfaceid": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1585.565571] env[62814]: DEBUG nova.compute.manager [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Start spawning the instance on the hypervisor. {{(pid=62814) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2645}} [ 1585.566153] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1585.566415] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1585.566576] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1585.566755] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1585.566900] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1585.567054] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1585.567702] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1585.567702] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1585.567702] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1585.567702] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1585.567909] env[62814]: DEBUG nova.virt.hardware [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1585.568801] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f332d91e-ca28-4cc4-ad47-2ddd399a9e7b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1585.577448] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-714529b9-1a15-4b77-be67-15dc9ce0f6b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1585.776472] env[62814]: DEBUG oslo_concurrency.lockutils [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1585.776807] env[62814]: DEBUG nova.compute.manager [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Instance network_info: |[{"id": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "address": "fa:16:3e:1e:b4:10", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0fba9cd-6b", "ovs_interfaceid": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62814) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2003}} [ 1585.777253] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:b4:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '75ffb921-796a-40fe-9662-d3fc01547dcb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0fba9cd-6bc1-41dd-902b-4e5546433a59', 'vif_model': 'vmxnet3'}] {{(pid=62814) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1585.784631] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1585.784836] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Creating VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1585.785062] env[62814]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e786d2f-0388-44ed-a843-597f071bb008 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1585.805366] env[62814]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1585.805366] env[62814]: value = "task-4294995" [ 1585.805366] env[62814]: _type = "Task" [ 1585.805366] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1585.813112] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294995, 'name': CreateVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1586.049914] env[62814]: DEBUG nova.compute.manager [req-5e93cb15-f124-4c16-a80c-bb03b7853ee4 req-0f74f83c-a1ec-4f35-bfc8-7f778ee311d6 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Received event network-changed-c0fba9cd-6bc1-41dd-902b-4e5546433a59 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1586.049914] env[62814]: DEBUG nova.compute.manager [req-5e93cb15-f124-4c16-a80c-bb03b7853ee4 req-0f74f83c-a1ec-4f35-bfc8-7f778ee311d6 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Refreshing instance network info cache due to event network-changed-c0fba9cd-6bc1-41dd-902b-4e5546433a59. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1586.050167] env[62814]: DEBUG oslo_concurrency.lockutils [req-5e93cb15-f124-4c16-a80c-bb03b7853ee4 req-0f74f83c-a1ec-4f35-bfc8-7f778ee311d6 service nova] Acquiring lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1586.050261] env[62814]: DEBUG oslo_concurrency.lockutils [req-5e93cb15-f124-4c16-a80c-bb03b7853ee4 req-0f74f83c-a1ec-4f35-bfc8-7f778ee311d6 service nova] Acquired lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1586.050411] env[62814]: DEBUG nova.network.neutron [req-5e93cb15-f124-4c16-a80c-bb03b7853ee4 req-0f74f83c-a1ec-4f35-bfc8-7f778ee311d6 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Refreshing network info cache for port c0fba9cd-6bc1-41dd-902b-4e5546433a59 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1586.316440] env[62814]: DEBUG oslo_vmware.api [-] Task: {'id': task-4294995, 'name': CreateVM_Task, 'duration_secs': 0.321617} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1586.316798] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Created VM on the ESX host {{(pid=62814) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1586.317255] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '0761804a-6ad8-4c8f-9444-34ba8ba1a4a5', 'mount_device': '/dev/sda', 'device_type': None, 'delete_on_termination': True, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845895', 'volume_id': '16230161-766e-474b-9d7b-4bada573fe8b', 'name': 'volume-16230161-766e-474b-9d7b-4bada573fe8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3', 'attached_at': '', 'detached_at': '', 'volume_id': '16230161-766e-474b-9d7b-4bada573fe8b', 'serial': '16230161-766e-474b-9d7b-4bada573fe8b'}, 'guest_format': None, 'disk_bus': None, 'volume_type': None}], 'swap': None} {{(pid=62814) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1586.317464] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Root volume attach. Driver type: vmdk {{(pid=62814) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1586.318233] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0ccc44-7e14-458c-8874-d2b6ae954440 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1586.325705] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9084bce2-fd42-4c69-a909-6fa1ff9caa2b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1586.331434] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0694306-6a86-47f5-8b78-44446fca2bfd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1586.336971] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-886c10a0-4873-42df-bdb2-ae201ca957b1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1586.344039] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1586.344039] env[62814]: value = "task-4294996" [ 1586.344039] env[62814]: _type = "Task" [ 1586.344039] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1586.351176] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294996, 'name': RelocateVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1586.734095] env[62814]: DEBUG nova.network.neutron [req-5e93cb15-f124-4c16-a80c-bb03b7853ee4 req-0f74f83c-a1ec-4f35-bfc8-7f778ee311d6 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updated VIF entry in instance network info cache for port c0fba9cd-6bc1-41dd-902b-4e5546433a59. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1586.734546] env[62814]: DEBUG nova.network.neutron [req-5e93cb15-f124-4c16-a80c-bb03b7853ee4 req-0f74f83c-a1ec-4f35-bfc8-7f778ee311d6 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance_info_cache with network_info: [{"id": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "address": "fa:16:3e:1e:b4:10", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0fba9cd-6b", "ovs_interfaceid": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1586.855607] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294996, 'name': RelocateVM_Task, 'duration_secs': 0.382931} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1586.855854] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Volume attach. Driver type: vmdk {{(pid=62814) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1586.856067] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845895', 'volume_id': '16230161-766e-474b-9d7b-4bada573fe8b', 'name': 'volume-16230161-766e-474b-9d7b-4bada573fe8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3', 'attached_at': '', 'detached_at': '', 'volume_id': '16230161-766e-474b-9d7b-4bada573fe8b', 'serial': '16230161-766e-474b-9d7b-4bada573fe8b'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1586.856790] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39924ba7-c256-4afe-abaf-5f19ddc998fa {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1586.872042] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fd12769-5c00-4725-911d-d2668811096c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1586.893358] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Reconfiguring VM instance instance-0000007f to attach disk [datastore2] volume-16230161-766e-474b-9d7b-4bada573fe8b/volume-16230161-766e-474b-9d7b-4bada573fe8b.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1586.893577] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2bea840-2d9c-419b-9714-1e52c30ef63e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1586.912546] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1586.912546] env[62814]: value = "task-4294997" [ 1586.912546] env[62814]: _type = "Task" [ 1586.912546] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1586.920034] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294997, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1587.238230] env[62814]: DEBUG oslo_concurrency.lockutils [req-5e93cb15-f124-4c16-a80c-bb03b7853ee4 req-0f74f83c-a1ec-4f35-bfc8-7f778ee311d6 service nova] Releasing lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1587.423167] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294997, 'name': ReconfigVM_Task, 'duration_secs': 0.2656} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1587.423532] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Reconfigured VM instance instance-0000007f to attach disk [datastore2] volume-16230161-766e-474b-9d7b-4bada573fe8b/volume-16230161-766e-474b-9d7b-4bada573fe8b.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1587.428120] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7b9b4ff9-f524-445b-9aac-f13b939bb270 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1587.443926] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1587.443926] env[62814]: value = "task-4294998" [ 1587.443926] env[62814]: _type = "Task" [ 1587.443926] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1587.452170] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294998, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1587.953776] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294998, 'name': ReconfigVM_Task, 'duration_secs': 0.139912} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1587.954045] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845895', 'volume_id': '16230161-766e-474b-9d7b-4bada573fe8b', 'name': 'volume-16230161-766e-474b-9d7b-4bada573fe8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3', 'attached_at': '', 'detached_at': '', 'volume_id': '16230161-766e-474b-9d7b-4bada573fe8b', 'serial': '16230161-766e-474b-9d7b-4bada573fe8b'} {{(pid=62814) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1587.954586] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c342af8-cea0-4362-a4ca-ae4cbd148467 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1587.961161] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1587.961161] env[62814]: value = "task-4294999" [ 1587.961161] env[62814]: _type = "Task" [ 1587.961161] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1587.968812] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294999, 'name': Rename_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1588.471421] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4294999, 'name': Rename_Task, 'duration_secs': 0.12365} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1588.471819] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1588.471926] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-88a4e8f8-9531-4105-a267-dee80b34e0bb {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1588.478643] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1588.478643] env[62814]: value = "task-4295000" [ 1588.478643] env[62814]: _type = "Task" [ 1588.478643] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1588.487156] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295000, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1588.988539] env[62814]: DEBUG oslo_vmware.api [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295000, 'name': PowerOnVM_Task, 'duration_secs': 0.445462} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1588.988818] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1588.989083] env[62814]: INFO nova.compute.manager [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Took 3.42 seconds to spawn the instance on the hypervisor. [ 1588.989277] env[62814]: DEBUG nova.compute.manager [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Checking state {{(pid=62814) _get_power_state /opt/stack/nova/nova/compute/manager.py:1798}} [ 1588.990144] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fea3755-80ef-45e0-98ce-90442c84bbef {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1589.510074] env[62814]: INFO nova.compute.manager [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Took 9.83 seconds to build instance. [ 1590.012148] env[62814]: DEBUG oslo_concurrency.lockutils [None req-df042e77-c9a7-462f-b416-c09546cc0b8a tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.342s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1590.731099] env[62814]: DEBUG nova.compute.manager [req-b9751c8c-5779-406e-b745-a1d5ddeeba47 req-c15151be-497b-4214-908d-5561ba2a5cbf service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Received event network-changed-cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1590.731362] env[62814]: DEBUG nova.compute.manager [req-b9751c8c-5779-406e-b745-a1d5ddeeba47 req-c15151be-497b-4214-908d-5561ba2a5cbf service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Refreshing instance network info cache due to event network-changed-cfcbae8d-a6f2-4583-9b46-101a3b73ac65. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1590.731560] env[62814]: DEBUG oslo_concurrency.lockutils [req-b9751c8c-5779-406e-b745-a1d5ddeeba47 req-c15151be-497b-4214-908d-5561ba2a5cbf service nova] Acquiring lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1590.731847] env[62814]: DEBUG oslo_concurrency.lockutils [req-b9751c8c-5779-406e-b745-a1d5ddeeba47 req-c15151be-497b-4214-908d-5561ba2a5cbf service nova] Acquired lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1590.731894] env[62814]: DEBUG nova.network.neutron [req-b9751c8c-5779-406e-b745-a1d5ddeeba47 req-c15151be-497b-4214-908d-5561ba2a5cbf service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Refreshing network info cache for port cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1591.427921] env[62814]: DEBUG nova.network.neutron [req-b9751c8c-5779-406e-b745-a1d5ddeeba47 req-c15151be-497b-4214-908d-5561ba2a5cbf service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Updated VIF entry in instance network info cache for port cfcbae8d-a6f2-4583-9b46-101a3b73ac65. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1591.428489] env[62814]: DEBUG nova.network.neutron [req-b9751c8c-5779-406e-b745-a1d5ddeeba47 req-c15151be-497b-4214-908d-5561ba2a5cbf service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Updating instance_info_cache with network_info: [{"id": "cfcbae8d-a6f2-4583-9b46-101a3b73ac65", "address": "fa:16:3e:30:f5:07", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcfcbae8d-a6", "ovs_interfaceid": "cfcbae8d-a6f2-4583-9b46-101a3b73ac65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1591.932187] env[62814]: DEBUG oslo_concurrency.lockutils [req-b9751c8c-5779-406e-b745-a1d5ddeeba47 req-c15151be-497b-4214-908d-5561ba2a5cbf service nova] Releasing lock "refresh_cache-711f9623-4c61-47da-b7e5-15a41487b3b1" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1592.221419] env[62814]: DEBUG nova.compute.manager [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Stashing vm_state: active {{(pid=62814) _prep_resize /opt/stack/nova/nova/compute/manager.py:6136}} [ 1592.741070] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1592.741352] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1592.757705] env[62814]: DEBUG nova.compute.manager [req-0161aa45-8a70-432b-b8be-3a967ae1bc2a req-8ce6521b-2292-4a33-9676-a75198d4ab10 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Received event network-changed-c0fba9cd-6bc1-41dd-902b-4e5546433a59 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1592.757861] env[62814]: DEBUG nova.compute.manager [req-0161aa45-8a70-432b-b8be-3a967ae1bc2a req-8ce6521b-2292-4a33-9676-a75198d4ab10 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Refreshing instance network info cache due to event network-changed-c0fba9cd-6bc1-41dd-902b-4e5546433a59. {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11773}} [ 1592.758088] env[62814]: DEBUG oslo_concurrency.lockutils [req-0161aa45-8a70-432b-b8be-3a967ae1bc2a req-8ce6521b-2292-4a33-9676-a75198d4ab10 service nova] Acquiring lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1592.758325] env[62814]: DEBUG oslo_concurrency.lockutils [req-0161aa45-8a70-432b-b8be-3a967ae1bc2a req-8ce6521b-2292-4a33-9676-a75198d4ab10 service nova] Acquired lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1592.758515] env[62814]: DEBUG nova.network.neutron [req-0161aa45-8a70-432b-b8be-3a967ae1bc2a req-8ce6521b-2292-4a33-9676-a75198d4ab10 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Refreshing network info cache for port c0fba9cd-6bc1-41dd-902b-4e5546433a59 {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1593.246498] env[62814]: INFO nova.compute.claims [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1593.507766] env[62814]: DEBUG nova.network.neutron [req-0161aa45-8a70-432b-b8be-3a967ae1bc2a req-8ce6521b-2292-4a33-9676-a75198d4ab10 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updated VIF entry in instance network info cache for port c0fba9cd-6bc1-41dd-902b-4e5546433a59. {{(pid=62814) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1593.508242] env[62814]: DEBUG nova.network.neutron [req-0161aa45-8a70-432b-b8be-3a967ae1bc2a req-8ce6521b-2292-4a33-9676-a75198d4ab10 service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance_info_cache with network_info: [{"id": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "address": "fa:16:3e:1e:b4:10", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0fba9cd-6b", "ovs_interfaceid": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1593.752704] env[62814]: INFO nova.compute.resource_tracker [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating resource usage from migration 01eaa7c8-1891-46fa-93dc-ad866938d389 [ 1593.819550] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea792204-c04e-42c0-9843-dce3f31c7154 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.827944] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09faef59-d4e8-484f-ba32-6f54466a30b7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.857733] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b09ee5-536d-4aba-9cef-84763976b997 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.865245] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0866f99a-a916-4069-83ff-bf1986909c77 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1593.878749] env[62814]: DEBUG nova.compute.provider_tree [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1594.011436] env[62814]: DEBUG oslo_concurrency.lockutils [req-0161aa45-8a70-432b-b8be-3a967ae1bc2a req-8ce6521b-2292-4a33-9676-a75198d4ab10 service nova] Releasing lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1594.381955] env[62814]: DEBUG nova.scheduler.client.report [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1594.886778] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.145s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1594.887049] env[62814]: INFO nova.compute.manager [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Migrating [ 1595.402758] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1595.403191] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1595.403191] env[62814]: DEBUG nova.network.neutron [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1596.094867] env[62814]: DEBUG nova.network.neutron [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance_info_cache with network_info: [{"id": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "address": "fa:16:3e:1e:b4:10", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0fba9cd-6b", "ovs_interfaceid": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1596.598123] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1598.114900] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c92cb69a-b761-4fd9-815f-a30c8aa133d9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1598.133884] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3' progress to 0 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1598.640222] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1598.640511] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-94e687d2-8d56-43eb-ac1e-e768e8ba7f66 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1598.651058] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1598.651058] env[62814]: value = "task-4295001" [ 1598.651058] env[62814]: _type = "Task" [ 1598.651058] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1598.658467] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295001, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1599.159959] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295001, 'name': PowerOffVM_Task, 'duration_secs': 0.19274} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1599.160295] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1599.160429] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3' progress to 17 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1599.668062] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:38Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1599.668062] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1599.668062] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1599.668062] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1599.668062] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1599.668471] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1599.668471] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1599.668609] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1599.668765] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1599.668927] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1599.669124] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1599.674061] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-09b7fbb4-bfd0-489e-8394-d72beb87857f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1599.690856] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1599.690856] env[62814]: value = "task-4295002" [ 1599.690856] env[62814]: _type = "Task" [ 1599.690856] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1599.699056] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295002, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1600.200748] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295002, 'name': ReconfigVM_Task, 'duration_secs': 0.147496} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1600.201107] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3' progress to 33 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1600.708076] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-06-21T05:32:36Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:567}} [ 1600.708383] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1600.708530] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image limits 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 1600.708728] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Flavor pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1600.708875] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Image pref 0:0:0 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 1600.709029] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62814) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:434}} [ 1600.709305] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:573}} [ 1600.709482] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:475}} [ 1600.709652] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Got 1 possible topologies {{(pid=62814) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:505}} [ 1600.709814] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:579}} [ 1600.709985] env[62814]: DEBUG nova.virt.hardware [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62814) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:581}} [ 1600.715504] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Reconfiguring VM instance instance-0000007f to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1600.715800] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9beb0592-a48a-495e-9ba2-caef4c1e3b93 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1600.734540] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1600.734540] env[62814]: value = "task-4295003" [ 1600.734540] env[62814]: _type = "Task" [ 1600.734540] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1600.742297] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295003, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1601.245140] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295003, 'name': ReconfigVM_Task, 'duration_secs': 0.155299} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1601.245517] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Reconfigured VM instance instance-0000007f to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1601.246313] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85bf5057-7043-4e57-af20-85344f5d1bc5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1601.269121] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Reconfiguring VM instance instance-0000007f to attach disk [datastore2] volume-16230161-766e-474b-9d7b-4bada573fe8b/volume-16230161-766e-474b-9d7b-4bada573fe8b.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1601.269800] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f4c3d17-4128-4aac-b3ea-76ad06ba83b2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1601.298014] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1601.298014] env[62814]: value = "task-4295004" [ 1601.298014] env[62814]: _type = "Task" [ 1601.298014] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1601.306918] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295004, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1601.807809] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1602.308458] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295004, 'name': ReconfigVM_Task, 'duration_secs': 0.79313} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1602.308857] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Reconfigured VM instance instance-0000007f to attach disk [datastore2] volume-16230161-766e-474b-9d7b-4bada573fe8b/volume-16230161-766e-474b-9d7b-4bada573fe8b.vmdk or device None with type thin {{(pid=62814) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1602.309025] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3' progress to 50 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1602.817026] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a30063-7333-41fc-9fd1-95c2d6138764 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1602.835864] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54839eea-0a7f-4381-a346-c51d0c153aa6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1602.854135] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3' progress to 67 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1604.497595] env[62814]: DEBUG nova.network.neutron [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Port c0fba9cd-6bc1-41dd-902b-4e5546433a59 binding to destination host cpu-1 is already ACTIVE {{(pid=62814) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1605.302729] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1605.302968] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1605.303149] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1605.303311] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1605.303463] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1605.303607] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1605.303751] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1605.303946] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 1605.304054] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1605.516087] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1605.516406] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1605.516507] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1605.808061] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1605.808061] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1605.808061] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1605.808061] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1605.808954] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bfc680f-15a6-4966-a707-6f7df5e45ab4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1605.817577] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9f0011-261b-401f-b337-a0cbf7ead48f {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1605.831055] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e170e947-c504-4f3a-9d46-09018331cb1b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1605.837106] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0bfcd80-99fa-42f7-adf7-0d6fd43dcaf0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1605.866998] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180559MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1605.867143] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1605.867355] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1606.556390] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1606.556622] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1606.556762] env[62814]: DEBUG nova.network.neutron [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1606.874222] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Applying migration context for instance a7491d42-9906-4a7c-9f3d-5e5dff3eddb3 as it has an incoming, in-progress migration 01eaa7c8-1891-46fa-93dc-ad866938d389. Migration status is post-migrating {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1606.874951] env[62814]: INFO nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating resource usage from migration 01eaa7c8-1891-46fa-93dc-ad866938d389 [ 1606.895491] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 711f9623-4c61-47da-b7e5-15a41487b3b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1606.895652] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 7099d3d1-c165-4600-b4ab-610161a9834d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1606.895781] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Migration 01eaa7c8-1891-46fa-93dc-ad866938d389 is active on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1606.895904] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance a7491d42-9906-4a7c-9f3d-5e5dff3eddb3 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1606.896091] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1606.896235] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=149GB used_disk=2GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1606.948869] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d37536e-4c54-4a45-a71b-c53c1f522ed3 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.956657] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1da94e9-eb7a-4e03-9e7e-4550a897871c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.986579] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3391f92-3dae-4297-aea3-9528c976f103 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1606.993459] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb7492c-a088-486e-aa00-6d9b802204d0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1607.006160] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1607.273297] env[62814]: DEBUG nova.network.neutron [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance_info_cache with network_info: [{"id": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "address": "fa:16:3e:1e:b4:10", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0fba9cd-6b", "ovs_interfaceid": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1607.509507] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1607.776351] env[62814]: DEBUG oslo_concurrency.lockutils [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1608.014638] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1608.014861] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.147s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1608.285572] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97661226-ffba-4e6c-889b-98c90021da98 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1608.300050] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-924c98e4-3b18-47d4-a56c-fec0543d0619 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.403221] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e6812d-26ef-4a04-a4d1-958aee641f8e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.422236] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973b2e14-2dd1-414d-a052-bb699b309341 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.428724] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3' progress to 83 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1609.935018] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Powering on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1609.935352] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4398b671-d292-41b6-8a19-a67fef15300e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1609.942823] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1609.942823] env[62814]: value = "task-4295005" [ 1609.942823] env[62814]: _type = "Task" [ 1609.942823] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1609.951825] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295005, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1610.454092] env[62814]: DEBUG oslo_vmware.api [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295005, 'name': PowerOnVM_Task, 'duration_secs': 0.408} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1610.454579] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Powered on the VM {{(pid=62814) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1610.454579] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-8895dbfc-828a-4572-9224-124718421413 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3' progress to 100 {{(pid=62814) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1612.791391] env[62814]: DEBUG oslo_concurrency.lockutils [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1612.793052] env[62814]: DEBUG oslo_concurrency.lockutils [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1612.793052] env[62814]: DEBUG nova.compute.manager [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Going to confirm migration 10 {{(pid=62814) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5250}} [ 1613.331571] env[62814]: DEBUG oslo_concurrency.lockutils [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1613.331964] env[62814]: DEBUG oslo_concurrency.lockutils [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquired lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 1613.332293] env[62814]: DEBUG nova.network.neutron [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Building network info cache for instance {{(pid=62814) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1613.332605] env[62814]: DEBUG nova.objects.instance [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lazy-loading 'info_cache' on Instance uuid a7491d42-9906-4a7c-9f3d-5e5dff3eddb3 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1614.561093] env[62814]: DEBUG nova.network.neutron [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance_info_cache with network_info: [{"id": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "address": "fa:16:3e:1e:b4:10", "network": {"id": "1b20c431-7b86-48b7-8cfb-389380434b9c", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-131061402-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.141", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "eb7d25c2161f40db8197b1a56517d7f0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "75ffb921-796a-40fe-9662-d3fc01547dcb", "external-id": "nsx-vlan-transportzone-331", "segmentation_id": 331, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0fba9cd-6b", "ovs_interfaceid": "c0fba9cd-6bc1-41dd-902b-4e5546433a59", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1615.063914] env[62814]: DEBUG oslo_concurrency.lockutils [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Releasing lock "refresh_cache-a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" {{(pid=62814) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 1615.064218] env[62814]: DEBUG nova.objects.instance [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lazy-loading 'migration_context' on Instance uuid a7491d42-9906-4a7c-9f3d-5e5dff3eddb3 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1615.567207] env[62814]: DEBUG nova.objects.base [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62814) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1615.568209] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df6bc7e-91c1-4d13-bb44-acb4e75da363 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1615.588322] env[62814]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91bedecd-00dc-4804-a329-c4cc63f2b697 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1615.593249] env[62814]: DEBUG oslo_vmware.api [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1615.593249] env[62814]: value = "session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ca9885-abc8-34fe-50fc-d9c3948c1c90" [ 1615.593249] env[62814]: _type = "Task" [ 1615.593249] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1615.600453] env[62814]: DEBUG oslo_vmware.api [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ca9885-abc8-34fe-50fc-d9c3948c1c90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1616.103785] env[62814]: DEBUG oslo_vmware.api [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': session[5213b259-a0d5-4270-44a4-1ed38c74b38c]52ca9885-abc8-34fe-50fc-d9c3948c1c90, 'name': SearchDatastore_Task, 'duration_secs': 0.01} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1616.104083] env[62814]: DEBUG oslo_concurrency.lockutils [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1616.104334] env[62814]: DEBUG oslo_concurrency.lockutils [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1616.669918] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2749220-b25e-430d-a02c-961f56834459 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1616.677212] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b20d2a2-c891-4159-a2ab-459db4c0303a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1616.706458] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c79941-696c-4f0f-8fba-ed96c6410b3a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1616.713636] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f43c598-0f2b-4fc5-b980-d163a3c94a68 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1616.727974] env[62814]: DEBUG nova.compute.provider_tree [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1617.231543] env[62814]: DEBUG nova.scheduler.client.report [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1618.242705] env[62814]: DEBUG oslo_concurrency.lockutils [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.138s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1618.807842] env[62814]: INFO nova.scheduler.client.report [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted allocation for migration 01eaa7c8-1891-46fa-93dc-ad866938d389 [ 1619.122609] env[62814]: INFO nova.compute.manager [None req-6f8d08d3-2558-419c-bad9-12a14e274889 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Get console output [ 1619.122842] env[62814]: WARNING nova.virt.vmwareapi.driver [None req-6f8d08d3-2558-419c-bad9-12a14e274889 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] The console log is missing. Check your VSPC configuration [ 1619.313158] env[62814]: DEBUG oslo_concurrency.lockutils [None req-70a0be62-6ae8-42cb-974d-0a34474c7774 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.521s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1646.109852] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1646.109852] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1646.109852] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1646.109852] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1646.110364] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1646.113583] env[62814]: INFO nova.compute.manager [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Terminating instance [ 1646.618048] env[62814]: DEBUG nova.compute.manager [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1646.618048] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1646.618273] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af5e34fa-7eb9-4f9f-848d-2fccdc0af75c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1646.625732] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1646.625732] env[62814]: value = "task-4295006" [ 1646.625732] env[62814]: _type = "Task" [ 1646.625732] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1646.635252] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295006, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1647.136188] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295006, 'name': PowerOffVM_Task, 'duration_secs': 0.169622} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1647.139046] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1647.139046] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Volume detach. Driver type: vmdk {{(pid=62814) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1647.139046] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845895', 'volume_id': '16230161-766e-474b-9d7b-4bada573fe8b', 'name': 'volume-16230161-766e-474b-9d7b-4bada573fe8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3', 'attached_at': '2025-06-21T05:50:57.000000', 'detached_at': '', 'volume_id': '16230161-766e-474b-9d7b-4bada573fe8b', 'serial': '16230161-766e-474b-9d7b-4bada573fe8b'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1647.139046] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f5674b-e8bc-45bd-9f98-9d0360efd264 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1647.156900] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e8e596-23bc-4c8a-8f54-21a068e67ab1 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1647.163091] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b85cb80-39b3-44eb-81fa-c4cdbfd37032 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1647.179993] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e35300-938a-4e7c-bcfd-d9517ec47309 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1647.193886] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] The volume has not been displaced from its original location: [datastore2] volume-16230161-766e-474b-9d7b-4bada573fe8b/volume-16230161-766e-474b-9d7b-4bada573fe8b.vmdk. No consolidation needed. {{(pid=62814) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1647.199114] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Reconfiguring VM instance instance-0000007f to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1647.199471] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3bc1b1d2-a261-4747-ab18-d2da19ab4b4d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1647.216202] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1647.216202] env[62814]: value = "task-4295007" [ 1647.216202] env[62814]: _type = "Task" [ 1647.216202] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1647.223908] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295007, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1647.726649] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295007, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1648.227165] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295007, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1648.728018] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295007, 'name': ReconfigVM_Task, 'duration_secs': 1.167189} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1648.728336] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Reconfigured VM instance instance-0000007f to detach disk 2000 {{(pid=62814) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1648.733150] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-58162d7a-faaa-4c2a-822d-f4163ca22efd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1648.747489] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1648.747489] env[62814]: value = "task-4295008" [ 1648.747489] env[62814]: _type = "Task" [ 1648.747489] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1648.755586] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295008, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1649.257381] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295008, 'name': ReconfigVM_Task, 'duration_secs': 0.096704} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1649.257723] env[62814]: DEBUG nova.virt.vmwareapi.volumeops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-845895', 'volume_id': '16230161-766e-474b-9d7b-4bada573fe8b', 'name': 'volume-16230161-766e-474b-9d7b-4bada573fe8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'attaching', 'instance': 'a7491d42-9906-4a7c-9f3d-5e5dff3eddb3', 'attached_at': '2025-06-21T05:50:57.000000', 'detached_at': '', 'volume_id': '16230161-766e-474b-9d7b-4bada573fe8b', 'serial': '16230161-766e-474b-9d7b-4bada573fe8b'} {{(pid=62814) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1649.257988] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1649.258720] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d25e83-e153-4c05-9896-72ca3ffb4b67 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1649.264812] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1649.265022] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-709234aa-7683-41a5-bef6-ab45351165a5 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1649.325558] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1649.325762] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1649.325945] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleting the datastore file [datastore2] a7491d42-9906-4a7c-9f3d-5e5dff3eddb3 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1649.326210] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-417bbcb5-11c9-4206-8161-426b17c9f4b0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1649.332713] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1649.332713] env[62814]: value = "task-4295010" [ 1649.332713] env[62814]: _type = "Task" [ 1649.332713] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1649.341962] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295010, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1649.842949] env[62814]: DEBUG oslo_vmware.api [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295010, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.080602} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1649.843284] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1649.843459] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1649.843629] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1649.843799] env[62814]: INFO nova.compute.manager [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Took 3.23 seconds to destroy the instance on the hypervisor. [ 1649.844114] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1649.844323] env[62814]: DEBUG nova.compute.manager [-] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1649.844417] env[62814]: DEBUG nova.network.neutron [-] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1650.311321] env[62814]: DEBUG nova.compute.manager [req-c7f062cf-2648-49c5-95d6-5ce994062634 req-c39fd936-9fe6-465e-b4c4-e2ec54b8658b service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Received event network-vif-deleted-c0fba9cd-6bc1-41dd-902b-4e5546433a59 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1650.311618] env[62814]: INFO nova.compute.manager [req-c7f062cf-2648-49c5-95d6-5ce994062634 req-c39fd936-9fe6-465e-b4c4-e2ec54b8658b service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Neutron deleted interface c0fba9cd-6bc1-41dd-902b-4e5546433a59; detaching it from the instance and deleting it from the info cache [ 1650.311808] env[62814]: DEBUG nova.network.neutron [req-c7f062cf-2648-49c5-95d6-5ce994062634 req-c39fd936-9fe6-465e-b4c4-e2ec54b8658b service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1650.795418] env[62814]: DEBUG nova.network.neutron [-] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1650.814635] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8267c162-0677-45f3-bf79-9d36d995b2a2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1650.825414] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ebcdead-f485-46b4-8022-b95d8e56c3a7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1650.851092] env[62814]: DEBUG nova.compute.manager [req-c7f062cf-2648-49c5-95d6-5ce994062634 req-c39fd936-9fe6-465e-b4c4-e2ec54b8658b service nova] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Detach interface failed, port_id=c0fba9cd-6bc1-41dd-902b-4e5546433a59, reason: Instance a7491d42-9906-4a7c-9f3d-5e5dff3eddb3 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1651.298705] env[62814]: INFO nova.compute.manager [-] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Took 1.45 seconds to deallocate network for instance. [ 1651.842616] env[62814]: INFO nova.compute.manager [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Took 0.54 seconds to detach 1 volumes for instance. [ 1651.844795] env[62814]: DEBUG nova.compute.manager [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: a7491d42-9906-4a7c-9f3d-5e5dff3eddb3] Deleting volume: 16230161-766e-474b-9d7b-4bada573fe8b {{(pid=62814) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3282}} [ 1652.384248] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1652.384541] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1652.384737] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1652.415410] env[62814]: INFO nova.scheduler.client.report [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted allocations for instance a7491d42-9906-4a7c-9f3d-5e5dff3eddb3 [ 1652.924137] env[62814]: DEBUG oslo_concurrency.lockutils [None req-a908b853-0a4b-409b-8f94-9f35716bfbea tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "a7491d42-9906-4a7c-9f3d-5e5dff3eddb3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.814s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1653.882696] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "7099d3d1-c165-4600-b4ab-610161a9834d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1653.882971] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "7099d3d1-c165-4600-b4ab-610161a9834d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1653.883266] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "7099d3d1-c165-4600-b4ab-610161a9834d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1653.883527] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "7099d3d1-c165-4600-b4ab-610161a9834d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1653.883745] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "7099d3d1-c165-4600-b4ab-610161a9834d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1653.885985] env[62814]: INFO nova.compute.manager [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Terminating instance [ 1654.389855] env[62814]: DEBUG nova.compute.manager [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1654.390235] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1654.391069] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c8c8ad1-2d58-4d2a-96b2-e6b1d32ac050 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1654.399102] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1654.399342] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1d3e45f-5ab0-4578-a160-ff4f3767124d {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1654.405724] env[62814]: DEBUG oslo_vmware.api [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1654.405724] env[62814]: value = "task-4295012" [ 1654.405724] env[62814]: _type = "Task" [ 1654.405724] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1654.413348] env[62814]: DEBUG oslo_vmware.api [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295012, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1654.915301] env[62814]: DEBUG oslo_vmware.api [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295012, 'name': PowerOffVM_Task, 'duration_secs': 0.180519} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1654.915516] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1654.915681] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1654.915927] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bfa93143-f2ff-4a33-a4e3-1473cacd0998 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1654.973764] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1654.974035] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1654.974182] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleting the datastore file [datastore2] 7099d3d1-c165-4600-b4ab-610161a9834d {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1654.974436] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9114c791-5fc6-4eef-b3ab-d6ace749f711 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1654.981590] env[62814]: DEBUG oslo_vmware.api [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1654.981590] env[62814]: value = "task-4295014" [ 1654.981590] env[62814]: _type = "Task" [ 1654.981590] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1654.989091] env[62814]: DEBUG oslo_vmware.api [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1655.492013] env[62814]: DEBUG oslo_vmware.api [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136968} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1655.492340] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1655.492440] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1655.492613] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1655.492783] env[62814]: INFO nova.compute.manager [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1655.493032] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1655.493254] env[62814]: DEBUG nova.compute.manager [-] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1655.493376] env[62814]: DEBUG nova.network.neutron [-] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1655.739378] env[62814]: DEBUG nova.compute.manager [req-81c77404-36ca-40a1-b7b3-3ff10069d273 req-10f70aeb-c282-4ad4-aeb7-32b115548f00 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Received event network-vif-deleted-9489e63a-3e31-4600-b530-7bde469fca61 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1655.739537] env[62814]: INFO nova.compute.manager [req-81c77404-36ca-40a1-b7b3-3ff10069d273 req-10f70aeb-c282-4ad4-aeb7-32b115548f00 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Neutron deleted interface 9489e63a-3e31-4600-b530-7bde469fca61; detaching it from the instance and deleting it from the info cache [ 1655.739743] env[62814]: DEBUG nova.network.neutron [req-81c77404-36ca-40a1-b7b3-3ff10069d273 req-10f70aeb-c282-4ad4-aeb7-32b115548f00 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1656.223308] env[62814]: DEBUG nova.network.neutron [-] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1656.242132] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f93b5d7f-3d21-443e-91ed-b5146a343f33 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1656.252543] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f92005-bdce-44b6-9482-42d00acc409e {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1656.277318] env[62814]: DEBUG nova.compute.manager [req-81c77404-36ca-40a1-b7b3-3ff10069d273 req-10f70aeb-c282-4ad4-aeb7-32b115548f00 service nova] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Detach interface failed, port_id=9489e63a-3e31-4600-b530-7bde469fca61, reason: Instance 7099d3d1-c165-4600-b4ab-610161a9834d could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1656.726601] env[62814]: INFO nova.compute.manager [-] [instance: 7099d3d1-c165-4600-b4ab-610161a9834d] Took 1.23 seconds to deallocate network for instance. [ 1657.233272] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1657.233571] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1657.233796] env[62814]: DEBUG nova.objects.instance [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lazy-loading 'resources' on Instance uuid 7099d3d1-c165-4600-b4ab-610161a9834d {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1657.781022] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7833d295-baae-4e13-ad4e-2ffc12a7452b {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1657.788202] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7675567-5518-4e27-b713-af024d601065 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1657.818249] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9d8f11-f635-4fa2-92d4-71882ba203f6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1657.825431] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03953584-9efd-42b1-9229-3b089d77e7d6 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1657.838384] env[62814]: DEBUG nova.compute.provider_tree [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1658.342221] env[62814]: DEBUG nova.scheduler.client.report [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1658.847870] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.614s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1658.866876] env[62814]: INFO nova.scheduler.client.report [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted allocations for instance 7099d3d1-c165-4600-b4ab-610161a9834d [ 1659.374931] env[62814]: DEBUG oslo_concurrency.lockutils [None req-4aaa8082-d67b-49bc-a497-8f79292ffe40 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "7099d3d1-c165-4600-b4ab-610161a9834d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.492s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1662.214064] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1662.214499] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1662.719565] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1662.719756] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1662.719900] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1662.720142] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1662.720332] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1662.720543] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1662.720690] env[62814]: DEBUG nova.compute.manager [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62814) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11184}} [ 1662.721144] env[62814]: DEBUG oslo_service.periodic_task [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62814) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1663.223449] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1663.223896] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1663.223957] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1663.224141] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62814) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1663.225243] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25292992-2fa6-4fa7-96c9-c206ef77cdc0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1663.233931] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c687786-04ed-4f07-861d-42a8f78a9c94 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1663.248371] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6152eaa-f574-4f6c-b5b9-f3f79c5f41b9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1663.254747] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c68b38d-ef89-4372-ae34-be47b2d0ad9c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1663.284036] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180320MB free_disk=145GB free_vcpus=48 pci_devices=None {{(pid=62814) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1663.284175] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1663.284386] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1663.652464] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "711f9623-4c61-47da-b7e5-15a41487b3b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1663.652730] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "711f9623-4c61-47da-b7e5-15a41487b3b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1663.652941] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "711f9623-4c61-47da-b7e5-15a41487b3b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1663.653173] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "711f9623-4c61-47da-b7e5-15a41487b3b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1663.653393] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "711f9623-4c61-47da-b7e5-15a41487b3b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1663.655617] env[62814]: INFO nova.compute.manager [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Terminating instance [ 1664.159847] env[62814]: DEBUG nova.compute.manager [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Start destroying the instance on the hypervisor. {{(pid=62814) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3164}} [ 1664.160115] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Destroying instance {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1664.161017] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5f947c-a051-4812-a629-17ee6b40ec1c {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.168843] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Powering off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1664.170067] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0cfd3cc7-e942-47f9-8903-46bd485215cd {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.175130] env[62814]: DEBUG oslo_vmware.api [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1664.175130] env[62814]: value = "task-4295016" [ 1664.175130] env[62814]: _type = "Task" [ 1664.175130] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1664.182798] env[62814]: DEBUG oslo_vmware.api [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295016, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1664.310038] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Instance 711f9623-4c61-47da-b7e5-15a41487b3b1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62814) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1664.310038] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1664.310397] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=149GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62814) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1664.339826] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a5dd155-f2b6-44ba-8bc9-9c3c5378e25a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.347691] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2ae5753-bde1-4836-988f-19bf4f0839c7 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.377273] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce78636-2fe8-42cd-a479-611e2202272a {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.384554] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847f394f-d3a3-41dd-a6e3-20361f105d49 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.397444] env[62814]: DEBUG nova.compute.provider_tree [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1664.686208] env[62814]: DEBUG oslo_vmware.api [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295016, 'name': PowerOffVM_Task, 'duration_secs': 0.205662} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1664.686537] env[62814]: DEBUG nova.virt.vmwareapi.vm_util [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Powered off the VM {{(pid=62814) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1664.686780] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Unregistering the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1664.687082] env[62814]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9ebcad07-97db-45b6-9a2f-421c01615108 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.751964] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Unregistered the VM {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1664.752163] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Deleting contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1664.752345] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleting the datastore file [datastore2] 711f9623-4c61-47da-b7e5-15a41487b3b1 {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1664.752597] env[62814]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b9ca20a6-2bdd-443c-ab17-e76a8263fb72 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1664.759106] env[62814]: DEBUG oslo_vmware.api [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for the task: (returnval){ [ 1664.759106] env[62814]: value = "task-4295018" [ 1664.759106] env[62814]: _type = "Task" [ 1664.759106] env[62814]: } to complete. {{(pid=62814) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1664.766446] env[62814]: DEBUG oslo_vmware.api [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295018, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1664.900809] env[62814]: DEBUG nova.scheduler.client.report [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1665.269140] env[62814]: DEBUG oslo_vmware.api [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Task: {'id': task-4295018, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131617} completed successfully. {{(pid=62814) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1665.269390] env[62814]: DEBUG nova.virt.vmwareapi.ds_util [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted the datastore file {{(pid=62814) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1665.269572] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Deleted contents of the VM from datastore datastore2 {{(pid=62814) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1665.269748] env[62814]: DEBUG nova.virt.vmwareapi.vmops [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Instance destroyed {{(pid=62814) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1665.269918] env[62814]: INFO nova.compute.manager [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1665.270216] env[62814]: DEBUG oslo.service.backend.eventlet.loopingcall [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62814) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/eventlet/loopingcall.py:436}} [ 1665.270401] env[62814]: DEBUG nova.compute.manager [-] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Deallocating network for instance {{(pid=62814) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2296}} [ 1665.270493] env[62814]: DEBUG nova.network.neutron [-] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] deallocate_for_instance() {{(pid=62814) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1665.405578] env[62814]: DEBUG nova.compute.resource_tracker [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62814) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1665.406082] env[62814]: DEBUG oslo_concurrency.lockutils [None req-502abaf4-3df7-47ef-b4e1-9b5a9706629a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.121s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1665.497816] env[62814]: DEBUG nova.compute.manager [req-db493ad1-9dbf-4eb7-bcce-019e7a6a0974 req-b450d473-4110-423d-bb51-1a4fe93acc20 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Received event network-vif-deleted-cfcbae8d-a6f2-4583-9b46-101a3b73ac65 {{(pid=62814) external_instance_event /opt/stack/nova/nova/compute/manager.py:11768}} [ 1665.498140] env[62814]: INFO nova.compute.manager [req-db493ad1-9dbf-4eb7-bcce-019e7a6a0974 req-b450d473-4110-423d-bb51-1a4fe93acc20 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Neutron deleted interface cfcbae8d-a6f2-4583-9b46-101a3b73ac65; detaching it from the instance and deleting it from the info cache [ 1665.498366] env[62814]: DEBUG nova.network.neutron [req-db493ad1-9dbf-4eb7-bcce-019e7a6a0974 req-b450d473-4110-423d-bb51-1a4fe93acc20 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1665.978445] env[62814]: DEBUG nova.network.neutron [-] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Updating instance_info_cache with network_info: [] {{(pid=62814) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1666.001283] env[62814]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f9fb1ca-94fc-4af8-b9af-f8a07e6a4bed {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1666.011443] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a335150e-dd08-4d20-957a-43f2058ea3a0 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1666.034990] env[62814]: DEBUG nova.compute.manager [req-db493ad1-9dbf-4eb7-bcce-019e7a6a0974 req-b450d473-4110-423d-bb51-1a4fe93acc20 service nova] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Detach interface failed, port_id=cfcbae8d-a6f2-4583-9b46-101a3b73ac65, reason: Instance 711f9623-4c61-47da-b7e5-15a41487b3b1 could not be found. {{(pid=62814) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11602}} [ 1666.481650] env[62814]: INFO nova.compute.manager [-] [instance: 711f9623-4c61-47da-b7e5-15a41487b3b1] Took 1.21 seconds to deallocate network for instance. [ 1666.989127] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 1666.989127] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 1666.989340] env[62814]: DEBUG nova.objects.instance [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lazy-loading 'resources' on Instance uuid 711f9623-4c61-47da-b7e5-15a41487b3b1 {{(pid=62814) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1667.526804] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33510646-b630-48e1-b105-3ae3b1296fc4 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1667.534540] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51eb39e3-5052-483f-98b3-6c4b2bafe835 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1667.565484] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2deccea-9951-47d1-ae26-74bfe680b4c2 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1667.572399] env[62814]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e513937-8167-4103-bbeb-3a495c821ef9 {{(pid=62814) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1667.584956] env[62814]: DEBUG nova.compute.provider_tree [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed in ProviderTree for provider: 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 {{(pid=62814) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1668.088031] env[62814]: DEBUG nova.scheduler.client.report [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Inventory has not changed for provider 7136a6f7-3927-4aa1-a4e9-7fcbd9976745 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 268, 'reserved': 0, 'min_unit': 1, 'max_unit': 145, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62814) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 1668.593543] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.604s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 1668.612908] env[62814]: INFO nova.scheduler.client.report [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Deleted allocations for instance 711f9623-4c61-47da-b7e5-15a41487b3b1 [ 1669.120865] env[62814]: DEBUG oslo_concurrency.lockutils [None req-174bf85f-2a08-431a-9208-d09b53963b24 tempest-ServerActionsTestOtherA-187063875 tempest-ServerActionsTestOtherA-187063875-project-member] Lock "711f9623-4c61-47da-b7e5-15a41487b3b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.468s {{(pid=62814) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}}